[ 525.982349] env[63355]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63355) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 525.982717] env[63355]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63355) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 525.982834] env[63355]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63355) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 525.983164] env[63355]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 526.079468] env[63355]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63355) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 526.090407] env[63355]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.011s {{(pid=63355) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 526.690345] env[63355]: INFO nova.virt.driver [None req-022b6211-3ba1-4983-86f7-846cb8d7cb76 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 526.762526] env[63355]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.762581] env[63355]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.762685] env[63355]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63355) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 529.844710] env[63355]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-8dcca8a0-41aa-4647-bf2f-d477310e6a92 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.860192] env[63355]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63355) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 529.860320] env[63355]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-b984bf0d-a781-40e2-8afc-122ffc82996a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.887231] env[63355]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 71a57. [ 529.887349] env[63355]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.125s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.887857] env[63355]: INFO nova.virt.vmwareapi.driver [None req-022b6211-3ba1-4983-86f7-846cb8d7cb76 None None] VMware vCenter version: 7.0.3 [ 529.891261] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c54b612-790c-40ec-8064-a51a5f33d0f6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.908313] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0f64fc-3854-4c3d-add2-e2ba54af38ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.914158] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5787b105-c793-407f-a22a-e8092ab6feb3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.920552] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45675be-862e-4548-94ae-8c6c9d6010c7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.934702] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d7e968-860c-40a2-b726-669c9566c6ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.940980] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a473f303-5410-4168-9e27-a233404c3637 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.971102] env[63355]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-9b886677-45f0-489f-9c8c-462b9c9ad10e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.976553] env[63355]: DEBUG nova.virt.vmwareapi.driver [None req-022b6211-3ba1-4983-86f7-846cb8d7cb76 None None] Extension org.openstack.compute already exists. {{(pid=63355) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 529.979362] env[63355]: INFO nova.compute.provider_config [None req-022b6211-3ba1-4983-86f7-846cb8d7cb76 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 530.483980] env[63355]: DEBUG nova.context [None req-022b6211-3ba1-4983-86f7-846cb8d7cb76 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),1e2c23ac-c75a-4b37-8332-b9f5b8c8986b(cell1) {{(pid=63355) load_cells /opt/stack/nova/nova/context.py:464}} [ 530.486091] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.486583] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.487412] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.488024] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Acquiring lock "1e2c23ac-c75a-4b37-8332-b9f5b8c8986b" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.490077] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Lock "1e2c23ac-c75a-4b37-8332-b9f5b8c8986b" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.490077] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Lock "1e2c23ac-c75a-4b37-8332-b9f5b8c8986b" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.509480] env[63355]: INFO dbcounter [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Registered counter for database nova_cell0 [ 530.517772] env[63355]: INFO dbcounter [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Registered counter for database nova_cell1 [ 530.521344] env[63355]: DEBUG oslo_db.sqlalchemy.engines [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63355) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 530.521863] env[63355]: DEBUG oslo_db.sqlalchemy.engines [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63355) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 530.526760] env[63355]: ERROR nova.db.main.api [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.526760] env[63355]: result = function(*args, **kwargs) [ 530.526760] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.526760] env[63355]: return func(*args, **kwargs) [ 530.526760] env[63355]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 530.526760] env[63355]: result = fn(*args, **kwargs) [ 530.526760] env[63355]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 530.526760] env[63355]: return f(*args, **kwargs) [ 530.526760] env[63355]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 530.526760] env[63355]: return db.service_get_minimum_version(context, binaries) [ 530.526760] env[63355]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 530.526760] env[63355]: _check_db_access() [ 530.526760] env[63355]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 530.526760] env[63355]: stacktrace = ''.join(traceback.format_stack()) [ 530.526760] env[63355]: [ 530.529364] env[63355]: ERROR nova.db.main.api [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.529364] env[63355]: result = function(*args, **kwargs) [ 530.529364] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.529364] env[63355]: return func(*args, **kwargs) [ 530.529364] env[63355]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 530.529364] env[63355]: result = fn(*args, **kwargs) [ 530.529364] env[63355]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 530.529364] env[63355]: return f(*args, **kwargs) [ 530.529364] env[63355]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 530.529364] env[63355]: return db.service_get_minimum_version(context, binaries) [ 530.529364] env[63355]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 530.529364] env[63355]: _check_db_access() [ 530.529364] env[63355]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 530.529364] env[63355]: stacktrace = ''.join(traceback.format_stack()) [ 530.529364] env[63355]: [ 530.529364] env[63355]: WARNING nova.objects.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Failed to get minimum service version for cell 1e2c23ac-c75a-4b37-8332-b9f5b8c8986b [ 530.529364] env[63355]: WARNING nova.objects.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 530.529364] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Acquiring lock "singleton_lock" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.529364] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Acquired lock "singleton_lock" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.530141] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Releasing lock "singleton_lock" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.530617] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Full set of CONF: {{(pid=63355) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 530.530925] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ******************************************************************************** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 530.534017] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Configuration options gathered from: {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 530.534017] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 530.534017] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 530.534017] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ================================================================================ {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 530.534017] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] allow_resize_to_same_host = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534017] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] arq_binding_timeout = 300 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534242] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] backdoor_port = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534242] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] backdoor_socket = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534242] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] block_device_allocate_retries = 60 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534242] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] block_device_allocate_retries_interval = 3 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534242] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cert = self.pem {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534242] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534242] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute_monitors = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534430] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] config_dir = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534430] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] config_drive_format = iso9660 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534430] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534430] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] config_source = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534430] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] console_host = devstack {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.534777] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] control_exchange = nova {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.535088] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cpu_allocation_ratio = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.535403] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] daemon = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.535715] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] debug = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.536085] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] default_access_ip_network_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.536423] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] default_availability_zone = nova {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.536726] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] default_ephemeral_format = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.537039] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] default_green_pool_size = 1000 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.537416] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.537771] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] default_schedule_zone = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.538076] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] disk_allocation_ratio = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.538393] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] enable_new_services = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.538704] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] enabled_apis = ['osapi_compute'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.539017] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] enabled_ssl_apis = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.539330] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] flat_injected = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.539660] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] force_config_drive = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.539973] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] force_raw_images = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543024] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] graceful_shutdown_timeout = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543024] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] heal_instance_info_cache_interval = 60 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543024] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] host = cpu-1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543024] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543024] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] initial_disk_allocation_ratio = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543024] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] initial_ram_allocation_ratio = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543024] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543245] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] instance_build_timeout = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543245] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] instance_delete_interval = 300 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543245] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] instance_format = [instance: %(uuid)s] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543245] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] instance_name_template = instance-%08x {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543245] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] instance_usage_audit = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543245] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] instance_usage_audit_period = month {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543245] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543426] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] instances_path = /opt/stack/data/nova/instances {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543426] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] internal_service_availability_zone = internal {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543684] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] key = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.543983] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] live_migration_retry_count = 30 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.544333] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] log_color = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.544643] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] log_config_append = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.544963] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.545277] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] log_dir = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.545569] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] log_file = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.545870] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] log_options = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.546193] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] log_rotate_interval = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.546649] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] log_rotate_interval_type = days {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.546959] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] log_rotation_type = none {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.547253] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.547515] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.547827] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] long_rpc_timeout = 1800 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] max_concurrent_builds = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] max_concurrent_live_migrations = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] max_concurrent_snapshots = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] max_local_block_devices = 3 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552255] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] max_logfile_count = 30 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552255] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] max_logfile_size_mb = 200 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552255] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] maximum_instance_delete_attempts = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552255] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] metadata_listen = 0.0.0.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552255] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] metadata_listen_port = 8775 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552255] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] metadata_workers = 2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552255] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] migrate_max_retries = -1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552438] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] mkisofs_cmd = genisoimage {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552438] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] my_block_storage_ip = 10.180.1.21 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552438] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] my_ip = 10.180.1.21 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552438] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] network_allocate_retries = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552438] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552438] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] osapi_compute_listen = 0.0.0.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552438] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] osapi_compute_listen_port = 8774 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552613] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] osapi_compute_unique_server_name_scope = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552613] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] osapi_compute_workers = 2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552613] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] password_length = 12 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.552613] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] periodic_enable = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.553101] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] periodic_fuzzy_delay = 60 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.553433] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] pointer_model = usbtablet {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.553763] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] preallocate_images = none {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.554097] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] publish_errors = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.554392] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] pybasedir = /opt/stack/nova {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.554684] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ram_allocation_ratio = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.554990] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] rate_limit_burst = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.555319] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] rate_limit_except_level = CRITICAL {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.555628] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] rate_limit_interval = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.555930] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] reboot_timeout = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.557016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] reclaim_instance_interval = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.557016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] record = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.557016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] reimage_timeout_per_gb = 60 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.557016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] report_interval = 120 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.557016] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] rescue_timeout = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.557456] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] reserved_host_cpus = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.557774] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] reserved_host_disk_mb = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.558106] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] reserved_host_memory_mb = 512 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.558408] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] reserved_huge_pages = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.558728] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] resize_confirm_window = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.559034] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] resize_fs_using_block_device = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.559346] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] resume_guests_state_on_host_boot = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.559649] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562064] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] rpc_response_timeout = 60 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562064] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] run_external_periodic_tasks = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562064] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] running_deleted_instance_action = reap {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562064] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] running_deleted_instance_poll_interval = 1800 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562064] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] running_deleted_instance_timeout = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562064] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler_instance_sync_interval = 120 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562064] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_down_time = 720 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562268] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] servicegroup_driver = db {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562268] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] shell_completion = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562268] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] shelved_offload_time = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562268] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] shelved_poll_interval = 3600 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562268] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] shutdown_timeout = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562268] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] source_is_ipv6 = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562268] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ssl_only = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.562823] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.563155] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] sync_power_state_interval = 600 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.563462] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] sync_power_state_pool_size = 1000 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.563767] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] syslog_log_facility = LOG_USER {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.564113] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] tempdir = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.564440] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] timeout_nbd = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.564761] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] transport_url = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.565082] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] update_resources_interval = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.566809] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] use_cow_images = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.566809] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] use_eventlog = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.566809] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] use_journal = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.566809] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] use_json = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.566809] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] use_rootwrap_daemon = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.566809] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] use_stderr = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.566809] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] use_syslog = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.567089] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vcpu_pin_set = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.567089] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plugging_is_fatal = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.567089] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plugging_timeout = 300 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.567089] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] virt_mkfs = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.567195] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] volume_usage_poll_interval = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.567374] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] watch_log_file = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.567502] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] web = /usr/share/spice-html5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 530.567688] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.567855] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.568025] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.568198] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_concurrency.disable_process_locking = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.568752] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.568947] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.569140] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.569317] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.569498] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.569741] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.569880] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.auth_strategy = keystone {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.570065] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.compute_link_prefix = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.570247] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.570426] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.dhcp_domain = novalocal {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.570595] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.enable_instance_password = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.570764] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.glance_link_prefix = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.570930] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.571116] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.571280] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.instance_list_per_project_cells = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.571443] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.list_records_by_skipping_down_cells = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.571617] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.local_metadata_per_cell = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.571810] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.max_limit = 1000 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.571982] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.metadata_cache_expiration = 15 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.572175] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.neutron_default_tenant_id = default {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.572348] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.response_validation = warn {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.572520] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.use_neutron_default_nets = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.572690] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.572853] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.573027] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.573204] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.573374] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.vendordata_dynamic_targets = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.573537] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.vendordata_jsonfile_path = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.573715] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.573912] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.backend = dogpile.cache.memcached {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.574117] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.backend_argument = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.574296] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.config_prefix = cache.oslo {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.574467] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.dead_timeout = 60.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.574632] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.debug_cache_backend = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.574797] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.enable_retry_client = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.574959] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.enable_socket_keepalive = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.575142] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.enabled = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.575307] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.enforce_fips_mode = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.575469] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.expiration_time = 600 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.575631] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.hashclient_retry_attempts = 2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.575798] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.hashclient_retry_delay = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.575961] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.memcache_dead_retry = 300 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.576132] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.memcache_password = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.576297] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.576459] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.576620] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.memcache_pool_maxsize = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.576784] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.576945] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.memcache_sasl_enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.577134] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.577300] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.memcache_socket_timeout = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.577458] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.memcache_username = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.577620] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.proxies = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.577784] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.redis_db = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.577942] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.redis_password = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.578151] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.redis_sentinel_service_name = mymaster {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.578334] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.578504] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.redis_server = localhost:6379 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.578668] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.redis_socket_timeout = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.578830] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.redis_username = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.578990] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.retry_attempts = 2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.579173] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.retry_delay = 0.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.579336] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.socket_keepalive_count = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.579506] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.socket_keepalive_idle = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.579684] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.socket_keepalive_interval = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.579849] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.tls_allowed_ciphers = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.580014] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.tls_cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.580183] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.tls_certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.580345] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.tls_enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.580501] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cache.tls_keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.580670] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.auth_section = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.580845] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.auth_type = password {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.581010] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.581193] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.catalog_info = volumev3::publicURL {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.581354] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.581516] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.581723] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.cross_az_attach = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.581895] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.debug = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.582068] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.endpoint_template = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.582237] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.http_retries = 3 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.582401] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.582557] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.582729] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.os_region_name = RegionOne {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.582895] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.583062] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cinder.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.583240] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.583400] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.cpu_dedicated_set = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.583557] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.cpu_shared_set = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.583720] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.image_type_exclude_list = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.583883] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.584080] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.max_concurrent_disk_ops = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.584249] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.max_disk_devices_to_attach = -1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.584412] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.584579] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.584798] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.resource_provider_association_refresh = 300 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.584978] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.585156] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.shutdown_retry_interval = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.585340] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.585516] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] conductor.workers = 2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.585694] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] console.allowed_origins = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.585857] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] console.ssl_ciphers = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.586038] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] console.ssl_minimum_version = default {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.586240] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] consoleauth.enforce_session_timeout = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.586452] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] consoleauth.token_ttl = 600 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.586634] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.586797] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.586986] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.587134] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.connect_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.587294] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.connect_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.587451] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.endpoint_override = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.587613] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.587772] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.587930] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.max_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.588098] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.min_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.588261] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.region_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.588417] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.retriable_status_codes = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.588573] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.service_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.588740] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.service_type = accelerator {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.588899] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.589067] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.status_code_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.589265] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.status_code_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.589379] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.589601] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.589927] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] cyborg.version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.590114] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.backend = sqlalchemy {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.590300] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.connection = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.590470] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.connection_debug = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.590643] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.connection_parameters = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.590839] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.connection_recycle_time = 3600 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.590974] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.connection_trace = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.591151] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.db_inc_retry_interval = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.591317] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.db_max_retries = 20 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.591478] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.db_max_retry_interval = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.591659] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.db_retry_interval = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.591834] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.max_overflow = 50 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.591998] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.max_pool_size = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.592176] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.max_retries = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.592342] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.592500] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.mysql_wsrep_sync_wait = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.592657] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.pool_timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.592817] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.retry_interval = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.592976] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.slave_connection = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.593150] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.sqlite_synchronous = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.593308] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] database.use_db_reconnect = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.593483] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.backend = sqlalchemy {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.593650] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.connection = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.593816] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.connection_debug = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.594008] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.connection_parameters = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.594191] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.connection_recycle_time = 3600 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.594354] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.connection_trace = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.594515] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.db_inc_retry_interval = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.594678] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.db_max_retries = 20 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.594840] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.db_max_retry_interval = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.594999] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.db_retry_interval = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.595173] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.max_overflow = 50 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.595333] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.max_pool_size = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.595493] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.max_retries = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.595661] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.595821] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.595977] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.pool_timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.596154] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.retry_interval = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.596311] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.slave_connection = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.596469] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] api_database.sqlite_synchronous = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.596643] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] devices.enabled_mdev_types = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.596819] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.596989] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ephemeral_storage_encryption.default_format = luks {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.597167] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ephemeral_storage_encryption.enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.597329] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.597497] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.api_servers = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.597661] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.597825] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.597987] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.598159] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.connect_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.598317] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.connect_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.598478] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.debug = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.598641] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.default_trusted_certificate_ids = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.598804] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.enable_certificate_validation = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.598964] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.enable_rbd_download = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.599136] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.endpoint_override = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.599303] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.599466] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.599627] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.max_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.599784] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.min_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.600037] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.num_retries = 3 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.600119] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.rbd_ceph_conf = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.600280] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.rbd_connect_timeout = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.600448] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.rbd_pool = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.600614] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.rbd_user = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.600774] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.region_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.600930] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.retriable_status_codes = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.601100] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.service_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.601288] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.service_type = image {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.601511] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.601712] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.status_code_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.601881] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.status_code_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.602050] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.602239] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.602404] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.verify_glance_signatures = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.602565] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] glance.version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.602733] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] guestfs.debug = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.602900] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] mks.enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.603269] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.603465] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] image_cache.manager_interval = 2400 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.603637] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] image_cache.precache_concurrency = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.603818] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] image_cache.remove_unused_base_images = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.604034] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.604229] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.604411] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] image_cache.subdirectory_name = _base {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.604590] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.api_max_retries = 60 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.604756] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.api_retry_interval = 2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.604916] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.auth_section = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.605091] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.auth_type = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.605254] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.605412] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.605574] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.605736] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.conductor_group = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.605896] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.connect_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.606069] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.connect_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.606231] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.endpoint_override = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.606393] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.606550] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.606709] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.max_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.606866] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.min_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.607040] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.peer_list = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.607204] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.region_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.607363] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.retriable_status_codes = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.607525] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.serial_console_state_timeout = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.607682] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.service_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.607853] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.service_type = baremetal {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.608020] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.shard = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.608185] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.608341] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.status_code_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.608497] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.status_code_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.608652] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.608833] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.608992] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ironic.version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.609187] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.609359] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] key_manager.fixed_key = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.609562] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.609741] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.barbican_api_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.609904] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.barbican_endpoint = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.610203] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.barbican_endpoint_type = public {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.610273] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.barbican_region_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.610396] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.610553] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.610715] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.610877] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.611043] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.611212] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.number_of_retries = 60 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.611374] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.retry_delay = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.611537] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.send_service_user_token = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.611721] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.611887] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.612058] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.verify_ssl = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.612219] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican.verify_ssl_path = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.612386] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican_service_user.auth_section = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.612549] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican_service_user.auth_type = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.612706] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican_service_user.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.612864] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican_service_user.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.613035] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican_service_user.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.613200] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican_service_user.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.613386] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican_service_user.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.613551] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican_service_user.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.613710] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] barbican_service_user.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.613878] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.approle_role_id = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.614045] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.approle_secret_id = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.614218] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.kv_mountpoint = secret {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.614376] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.kv_path = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.614539] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.kv_version = 2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.614695] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.namespace = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.614853] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.root_token_id = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.615020] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.ssl_ca_crt_file = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.615193] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.timeout = 60.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.615353] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.use_ssl = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.615521] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.615688] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.auth_section = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.615849] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.auth_type = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.616015] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.616183] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.616344] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.616501] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.connect_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.616657] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.connect_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.616813] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.endpoint_override = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.616970] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.617141] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.617295] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.max_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.617449] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.min_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.617603] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.region_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.617761] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.retriable_status_codes = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.617915] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.service_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.618093] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.service_type = identity {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.618257] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.618413] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.status_code_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.618571] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.status_code_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.618726] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.618909] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.619080] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] keystone.version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.619282] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.connection_uri = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.619442] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.cpu_mode = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.619646] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.cpu_model_extra_flags = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.619807] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.cpu_models = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.620015] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.cpu_power_governor_high = performance {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.620199] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.cpu_power_governor_low = powersave {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.620365] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.cpu_power_management = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.620534] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.620698] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.device_detach_attempts = 8 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.620860] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.device_detach_timeout = 20 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.621033] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.disk_cachemodes = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.621198] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.disk_prefix = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.621360] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.enabled_perf_events = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.621528] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.file_backed_memory = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.621734] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.gid_maps = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.621903] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.hw_disk_discard = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.622072] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.hw_machine_type = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.622247] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.images_rbd_ceph_conf = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.622414] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.622576] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.622744] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.images_rbd_glance_store_name = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.622910] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.images_rbd_pool = rbd {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.623088] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.images_type = default {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.623248] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.images_volume_group = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.623408] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.inject_key = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.623568] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.inject_partition = -2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.623726] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.inject_password = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.623887] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.iscsi_iface = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.624083] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.iser_use_multipath = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.624253] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_bandwidth = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.624415] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.624574] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_downtime = 500 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.624734] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.624899] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.625072] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_inbound_addr = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.625238] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.625417] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_permit_post_copy = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.625581] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_scheme = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.625762] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_timeout_action = abort {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.625934] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_tunnelled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.626104] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_uri = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.626268] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.live_migration_with_native_tls = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.626425] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.max_queues = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.626587] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.626821] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.626984] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.nfs_mount_options = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.627290] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.627465] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.627631] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.num_iser_scan_tries = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.627795] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.num_memory_encrypted_guests = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.627957] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.628134] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.num_pcie_ports = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.628298] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.num_volume_scan_tries = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.628460] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.pmem_namespaces = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.628619] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.quobyte_client_cfg = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.628901] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.629085] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.rbd_connect_timeout = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.629253] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.629417] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.629598] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.rbd_secret_uuid = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.629767] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.rbd_user = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.629931] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.630115] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.remote_filesystem_transport = ssh {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.630279] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.rescue_image_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.630436] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.rescue_kernel_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.630591] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.rescue_ramdisk_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.630759] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.630916] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.rx_queue_size = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.631095] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.smbfs_mount_options = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.631370] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.631541] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.snapshot_compression = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.631732] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.snapshot_image_format = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.631955] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.632136] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.sparse_logical_volumes = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.632304] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.swtpm_enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.632478] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.swtpm_group = tss {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.632652] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.swtpm_user = tss {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.632919] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.sysinfo_serial = unique {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.632979] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.tb_cache_size = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.633137] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.tx_queue_size = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.633301] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.uid_maps = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.633463] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.use_virtio_for_bridges = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.633654] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.virt_type = kvm {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.633835] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.volume_clear = zero {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.634026] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.volume_clear_size = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.634205] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.volume_use_multipath = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.634364] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.vzstorage_cache_path = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.634533] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.634698] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.vzstorage_mount_group = qemu {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.634863] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.vzstorage_mount_opts = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.635041] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.635320] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.635498] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.vzstorage_mount_user = stack {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.635668] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.635845] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.auth_section = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.636025] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.auth_type = password {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.636194] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.636353] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.636515] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.636670] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.connect_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.636827] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.connect_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.636992] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.default_floating_pool = public {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.637164] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.endpoint_override = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.637326] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.extension_sync_interval = 600 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.637481] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.http_retries = 3 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.637639] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.637798] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.637954] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.max_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.638134] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.638292] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.min_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.638456] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.ovs_bridge = br-int {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.638619] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.physnets = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.638788] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.region_name = RegionOne {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.638944] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.retriable_status_codes = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.639125] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.service_metadata_proxy = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.639284] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.service_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.639449] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.service_type = network {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.639636] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.639808] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.status_code_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.639959] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.status_code_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.640131] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.640355] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.640465] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] neutron.version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.640660] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] notifications.bdms_in_notifications = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.640805] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] notifications.default_level = INFO {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.640983] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] notifications.notification_format = unversioned {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.641160] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] notifications.notify_on_state_change = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.641337] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.641513] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] pci.alias = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.641717] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] pci.device_spec = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.641889] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] pci.report_in_placement = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.642076] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.auth_section = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.642252] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.auth_type = password {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.642422] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.642594] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.642750] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.642923] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.643096] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.connect_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.643259] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.connect_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.643414] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.default_domain_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.643571] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.default_domain_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.643725] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.domain_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.643883] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.domain_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.644068] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.endpoint_override = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.644239] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.644397] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.644554] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.max_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.644708] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.min_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.644876] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.password = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.645041] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.project_domain_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.645210] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.project_domain_name = Default {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.645379] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.project_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.645548] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.project_name = service {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.645745] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.region_name = RegionOne {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.645916] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.retriable_status_codes = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.646114] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.service_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.646290] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.service_type = placement {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.646454] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.646613] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.status_code_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.646774] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.status_code_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.646932] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.system_scope = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.647102] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.647266] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.trust_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.647427] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.user_domain_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.647589] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.user_domain_name = Default {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.647750] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.user_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.647923] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.username = nova {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.648113] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.648276] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] placement.version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.648451] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.cores = 20 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.648614] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.count_usage_from_placement = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.648782] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.648951] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.injected_file_content_bytes = 10240 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.649131] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.injected_file_path_length = 255 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.649298] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.injected_files = 5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.649462] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.instances = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.649653] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.key_pairs = 100 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.649822] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.metadata_items = 128 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.649988] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.ram = 51200 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.650166] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.recheck_quota = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.650333] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.server_group_members = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.650506] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] quota.server_groups = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.650665] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.650829] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.651021] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler.image_metadata_prefilter = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.651157] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.651321] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler.max_attempts = 3 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.651482] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler.max_placement_results = 1000 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.651664] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.651837] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler.query_placement_for_image_type_support = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.651997] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.652181] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] scheduler.workers = 2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.652352] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.652523] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.652707] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.652876] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.653047] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.653220] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.653378] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.653567] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.653734] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.host_subset_size = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.653902] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.654091] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.654257] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.654421] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.isolated_hosts = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.654583] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.isolated_images = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.654744] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.654905] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.655078] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.655241] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.pci_in_placement = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.655400] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.655559] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.655758] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.655932] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.656108] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.656270] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.656429] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.track_instance_changes = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.656610] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.656786] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] metrics.required = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.656951] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] metrics.weight_multiplier = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.657126] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.657293] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] metrics.weight_setting = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.657608] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.657787] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] serial_console.enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.658034] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] serial_console.port_range = 10000:20000 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.658241] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.658415] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.658585] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] serial_console.serialproxy_port = 6083 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.658754] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_user.auth_section = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.658929] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_user.auth_type = password {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.659102] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_user.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.659266] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_user.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.659429] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_user.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.659590] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_user.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.659746] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_user.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.659914] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_user.send_service_user_token = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.660087] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_user.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.660262] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] service_user.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.660429] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.agent_enabled = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.660592] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.660904] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.661105] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.661277] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.html5proxy_port = 6082 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.661437] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.image_compression = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.661599] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.jpeg_compression = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.661771] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.playback_compression = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.661937] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.require_secure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.662115] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.server_listen = 127.0.0.1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.662284] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.662439] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.streaming_mode = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.662596] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] spice.zlib_compression = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.662758] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] upgrade_levels.baseapi = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.662924] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] upgrade_levels.compute = auto {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.663099] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] upgrade_levels.conductor = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.663256] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] upgrade_levels.scheduler = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.663418] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vendordata_dynamic_auth.auth_section = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.663579] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vendordata_dynamic_auth.auth_type = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.663734] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vendordata_dynamic_auth.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.663893] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vendordata_dynamic_auth.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.664093] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.664261] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vendordata_dynamic_auth.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.664419] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vendordata_dynamic_auth.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.664582] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.664738] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vendordata_dynamic_auth.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.664912] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.api_retry_count = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.665082] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.ca_file = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.665253] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.cache_prefix = devstack-image-cache {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.665417] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.cluster_name = testcl1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.665579] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.connection_pool_size = 10 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.665734] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.console_delay_seconds = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.665898] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.datastore_regex = ^datastore.* {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.666113] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.666289] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.host_password = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.666454] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.host_port = 443 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.666619] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.host_username = administrator@vsphere.local {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.666782] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.insecure = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.666939] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.integration_bridge = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.667109] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.maximum_objects = 100 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.667267] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.pbm_default_policy = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.667425] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.pbm_enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.667579] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.pbm_wsdl_location = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.667812] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.667940] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.serial_port_proxy_uri = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.668112] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.serial_port_service_uri = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.668277] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.task_poll_interval = 0.5 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.668443] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.use_linked_clone = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.668609] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.vnc_keymap = en-us {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.668772] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.vnc_port = 5900 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.668933] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vmware.vnc_port_total = 10000 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.669128] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vnc.auth_schemes = ['none'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.669303] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vnc.enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.669596] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.669782] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.669952] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vnc.novncproxy_port = 6080 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.670204] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vnc.server_listen = 127.0.0.1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.670343] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.670505] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vnc.vencrypt_ca_certs = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.670664] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vnc.vencrypt_client_cert = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.670823] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vnc.vencrypt_client_key = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.671009] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.671179] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.disable_deep_image_inspection = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.671339] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.671499] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.671682] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.671854] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.disable_rootwrap = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.672020] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.enable_numa_live_migration = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.672187] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.672348] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.672507] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.672668] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.libvirt_disable_apic = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.672826] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.672987] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.673162] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.673322] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.673481] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.673640] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.673800] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.673984] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.674174] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.674342] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.674525] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.674693] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.client_socket_timeout = 900 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.674862] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.default_pool_size = 1000 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.675037] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.keep_alive = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.675209] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.max_header_line = 16384 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.675369] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.secure_proxy_ssl_header = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.675530] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.ssl_ca_file = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.675688] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.ssl_cert_file = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.675849] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.ssl_key_file = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.676025] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.tcp_keepidle = 600 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.676204] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.676370] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] zvm.ca_file = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.676531] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] zvm.cloud_connector_url = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.676812] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.676988] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] zvm.reachable_timeout = 300 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.677182] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_policy.enforce_new_defaults = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.677564] env[63355]: WARNING oslo_config.cfg [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 530.677748] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_policy.enforce_scope = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.677928] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_policy.policy_default_rule = default {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.678119] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.678297] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_policy.policy_file = policy.yaml {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.678471] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.678633] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.678796] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.678956] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.679136] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.679308] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.679481] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.679684] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.connection_string = messaging:// {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.679856] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.enabled = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.680038] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.es_doc_type = notification {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.680209] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.es_scroll_size = 10000 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.680377] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.es_scroll_time = 2m {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.680538] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.filter_error_trace = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.680736] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.hmac_keys = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.680866] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.sentinel_service_name = mymaster {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.681047] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.socket_timeout = 0.1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.681208] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.trace_requests = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.681367] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler.trace_sqlalchemy = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.681545] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler_jaeger.process_tags = {} {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.681728] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler_jaeger.service_name_prefix = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.681895] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] profiler_otlp.service_name_prefix = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.682084] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] remote_debug.host = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.682264] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] remote_debug.port = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.682444] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.682611] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.682775] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.682937] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.683108] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.683273] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.683432] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.683591] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.683752] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.683921] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.684092] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.684267] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.684433] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.684601] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.684771] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.684938] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.685109] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.685284] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.685447] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.685606] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.685775] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.685939] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.686111] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.686281] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.686441] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.686599] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.686759] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.686918] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.687092] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.687259] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.ssl = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.687432] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.687603] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.687765] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.687935] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.688117] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.ssl_version = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.688282] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.688521] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.688696] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_notifications.retry = -1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.688882] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.689069] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_messaging_notifications.transport_url = **** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.689246] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.auth_section = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.689408] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.auth_type = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.689588] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.cafile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.689758] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.certfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.689923] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.collect_timing = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.690097] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.connect_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.690262] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.connect_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.690440] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.endpoint_id = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.690574] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.endpoint_override = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.690732] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.insecure = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.690895] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.keyfile = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.691054] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.max_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.691220] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.min_version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.691376] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.region_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.691549] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.retriable_status_codes = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.691751] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.service_name = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.691907] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.service_type = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.692084] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.split_loggers = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.692248] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.status_code_retries = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.692406] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.status_code_retry_delay = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.692562] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.timeout = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.692721] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.valid_interfaces = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.692879] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_limit.version = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.693054] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_reports.file_event_handler = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.693223] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.693383] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] oslo_reports.log_dir = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.693554] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.693712] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.693872] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.694074] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.694248] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.694408] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.694577] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.694736] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_ovs_privileged.group = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.694893] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.695069] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.695234] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.695394] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] vif_plug_ovs_privileged.user = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.695562] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_linux_bridge.flat_interface = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.695737] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.695911] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.696091] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.696267] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.696433] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.696598] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.696763] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.696939] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.697121] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_ovs.isolate_vif = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.697292] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.697457] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.697624] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.697792] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_ovs.ovsdb_interface = native {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.697953] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] os_vif_ovs.per_port_bridge = False {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.698138] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] privsep_osbrick.capabilities = [21] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.698299] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] privsep_osbrick.group = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.698455] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] privsep_osbrick.helper_command = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.698620] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.698785] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.698940] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] privsep_osbrick.user = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.699129] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.699289] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] nova_sys_admin.group = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.699444] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] nova_sys_admin.helper_command = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.699630] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.699800] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.699956] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] nova_sys_admin.user = None {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 530.700100] env[63355]: DEBUG oslo_service.service [None req-af97d50b-66b1-4d10-a2fa-fcf4fcfbf7ef None None] ******************************************************************************** {{(pid=63355) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 530.700585] env[63355]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 531.203523] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Getting list of instances from cluster (obj){ [ 531.203523] env[63355]: value = "domain-c8" [ 531.203523] env[63355]: _type = "ClusterComputeResource" [ 531.203523] env[63355]: } {{(pid=63355) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 531.204735] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acea6ff5-ef8b-47bc-acf0-aeb3e9b74f36 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.213751] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Got total of 0 instances {{(pid=63355) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 531.214373] env[63355]: WARNING nova.virt.vmwareapi.driver [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 531.214795] env[63355]: INFO nova.virt.node [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Generated node identity 47abb610-db7e-4770-911d-187dd075ef8b [ 531.215043] env[63355]: INFO nova.virt.node [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Wrote node identity 47abb610-db7e-4770-911d-187dd075ef8b to /opt/stack/data/n-cpu-1/compute_id [ 531.718640] env[63355]: WARNING nova.compute.manager [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Compute nodes ['47abb610-db7e-4770-911d-187dd075ef8b'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 532.724579] env[63355]: INFO nova.compute.manager [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 533.729288] env[63355]: WARNING nova.compute.manager [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 533.729715] env[63355]: DEBUG oslo_concurrency.lockutils [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.729758] env[63355]: DEBUG oslo_concurrency.lockutils [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.729921] env[63355]: DEBUG oslo_concurrency.lockutils [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.730130] env[63355]: DEBUG nova.compute.resource_tracker [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 533.731046] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe57a95d-2617-4387-b0b6-66ac44806b4e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.739326] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01231a83-7ebf-4abc-bd1e-c95eecb1a2d3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.752816] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70c132d-b209-4e2e-b623-86bdf0bd70a0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.759276] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22750af9-bd58-402b-b987-38d6633d2a7d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.788379] env[63355]: DEBUG nova.compute.resource_tracker [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181355MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 533.788537] env[63355]: DEBUG oslo_concurrency.lockutils [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.788716] env[63355]: DEBUG oslo_concurrency.lockutils [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.290995] env[63355]: WARNING nova.compute.resource_tracker [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] No compute node record for cpu-1:47abb610-db7e-4770-911d-187dd075ef8b: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 47abb610-db7e-4770-911d-187dd075ef8b could not be found. [ 534.794563] env[63355]: INFO nova.compute.resource_tracker [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 47abb610-db7e-4770-911d-187dd075ef8b [ 536.303451] env[63355]: DEBUG nova.compute.resource_tracker [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 536.303804] env[63355]: DEBUG nova.compute.resource_tracker [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 536.452316] env[63355]: INFO nova.scheduler.client.report [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] [req-501c910f-2bcd-495d-b078-2a7dd673cc0d] Created resource provider record via placement API for resource provider with UUID 47abb610-db7e-4770-911d-187dd075ef8b and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 536.469153] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992b434c-9932-4728-9a89-f6682bd27383 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.476703] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5b440e-6cff-43e1-90fa-14a2c6323370 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.506125] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ddc1ea-2b7b-415e-bee4-6d90ef1842ba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.513237] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9feb9381-5b21-4d71-8ed7-a2930e479224 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.525992] env[63355]: DEBUG nova.compute.provider_tree [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 537.064241] env[63355]: DEBUG nova.scheduler.client.report [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Updated inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 537.064491] env[63355]: DEBUG nova.compute.provider_tree [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Updating resource provider 47abb610-db7e-4770-911d-187dd075ef8b generation from 0 to 1 during operation: update_inventory {{(pid=63355) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 537.064631] env[63355]: DEBUG nova.compute.provider_tree [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 537.113554] env[63355]: DEBUG nova.compute.provider_tree [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Updating resource provider 47abb610-db7e-4770-911d-187dd075ef8b generation from 1 to 2 during operation: update_traits {{(pid=63355) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 537.618358] env[63355]: DEBUG nova.compute.resource_tracker [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 537.618800] env[63355]: DEBUG oslo_concurrency.lockutils [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.830s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.618800] env[63355]: DEBUG nova.service [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Creating RPC server for service compute {{(pid=63355) start /opt/stack/nova/nova/service.py:186}} [ 537.632647] env[63355]: DEBUG nova.service [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] Join ServiceGroup membership for this service compute {{(pid=63355) start /opt/stack/nova/nova/service.py:203}} [ 537.632869] env[63355]: DEBUG nova.servicegroup.drivers.db [None req-629d572e-b38c-4f3d-beb5-b9e2bce19d47 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63355) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 574.390573] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Acquiring lock "e86f2c82-d544-4d73-bcc4-f55edf56cab3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.390900] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Lock "e86f2c82-d544-4d73-bcc4-f55edf56cab3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.893381] env[63355]: DEBUG nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 575.437936] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.438974] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.440107] env[63355]: INFO nova.compute.claims [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.510147] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da09328c-cccc-4734-9153-fe3cd6bd8cc5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.522090] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a9da09-67c2-4312-95b1-139bf83c7f5b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.555419] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1b1fbc-8f3a-42e2-a03b-dfee71f213d7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.563240] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef44ef9-02fc-4f91-8a3d-2086f9681573 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.577882] env[63355]: DEBUG nova.compute.provider_tree [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.081094] env[63355]: DEBUG nova.scheduler.client.report [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.288459] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Acquiring lock "35bd361f-fda1-4121-9f4c-61d3219f9202" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.288548] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Lock "35bd361f-fda1-4121-9f4c-61d3219f9202" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.587273] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.149s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.587934] env[63355]: DEBUG nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.792290] env[63355]: DEBUG nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 578.094784] env[63355]: DEBUG nova.compute.utils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.097027] env[63355]: DEBUG nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 578.097027] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 578.332961] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.332961] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.334694] env[63355]: INFO nova.compute.claims [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 578.383448] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquiring lock "b3d31162-5851-466f-b44c-d15e69c76bdd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.383711] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "b3d31162-5851-466f-b44c-d15e69c76bdd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.550766] env[63355]: DEBUG nova.policy [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5632ef27cab247dc8d6547123736b270', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5b4933a63a804241bc67c05bc1497f70', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 578.604976] env[63355]: DEBUG nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.886413] env[63355]: DEBUG nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.414902] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.431652] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7fde97-81e1-45c1-939d-2a1ec4c700ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.442413] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172f9da1-1469-45e3-a3eb-7d8b02fd9a84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.485364] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d3e22a-9580-4435-9c57-e8b3fe58a6dc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.495256] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7747ca-85e3-48b5-885f-4bd5261e50fb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.512619] env[63355]: DEBUG nova.compute.provider_tree [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.529194] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Successfully created port: 2d8e87a1-a834-438b-a0ca-187f648f9236 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.616085] env[63355]: DEBUG nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.637060] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Acquiring lock "d08c461b-c2ba-4484-a0aa-746d726868a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.637289] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Lock "d08c461b-c2ba-4484-a0aa-746d726868a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.656185] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.656709] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.656709] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.657105] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.657105] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.658407] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.658407] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.658546] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.658949] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.659150] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.659325] env[63355]: DEBUG nova.virt.hardware [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.660927] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64cfb72b-41d4-44e2-a8b9-30c5d1b94377 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.676166] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6235e540-4565-48f5-ab6e-1266b42c7807 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.699529] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98bdb2fd-1ce6-4b00-a73e-840b3411ea45 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.017171] env[63355]: DEBUG nova.scheduler.client.report [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.141399] env[63355]: DEBUG nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.526889] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.194s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.527513] env[63355]: DEBUG nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 580.530123] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.115s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.533234] env[63355]: INFO nova.compute.claims [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 580.678321] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.037121] env[63355]: DEBUG nova.compute.utils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 581.040982] env[63355]: DEBUG nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 581.040982] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 581.222137] env[63355]: DEBUG nova.policy [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbda037cf2b94e9880171d6eecb93354', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f0c3f2d18b241d5beee0ed80be7fa8e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 581.546543] env[63355]: DEBUG nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.685249] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bf4aa0-65c5-4e2b-8f31-bdd71ec83791 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.697581] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f06e5eb-00d0-413e-a065-2b067f3716ce {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.747263] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b210f28-c774-4226-b9e6-a1aa412f1eb0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.757216] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669c88aa-99e2-45ec-8032-2d27153a526a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.774034] env[63355]: DEBUG nova.compute.provider_tree [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.196767] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Successfully created port: 6038aaf0-9377-44f3-adc1-bb4c25210d18 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 582.284347] env[63355]: DEBUG nova.scheduler.client.report [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.488602] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquiring lock "059dc72c-dc91-4744-8efa-886aef99d5c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.488836] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "059dc72c-dc91-4744-8efa-886aef99d5c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.561228] env[63355]: DEBUG nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 582.594448] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.594448] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.594448] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.594650] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.594650] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.594650] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.594650] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.594785] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.595747] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.595747] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.595747] env[63355]: DEBUG nova.virt.hardware [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.596408] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1083f8-7c3c-40d1-a6e0-26240def03e3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.604878] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13dd80d4-c64d-456a-bda9-7df746687986 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.805468] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.805468] env[63355]: DEBUG nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 582.807816] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.130s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.812519] env[63355]: INFO nova.compute.claims [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.991610] env[63355]: DEBUG nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 583.253034] env[63355]: ERROR nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2d8e87a1-a834-438b-a0ca-187f648f9236, please check neutron logs for more information. [ 583.253034] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 583.253034] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.253034] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 583.253034] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.253034] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 583.253034] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.253034] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 583.253034] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.253034] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 583.253034] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.253034] env[63355]: ERROR nova.compute.manager raise self.value [ 583.253034] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.253034] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 583.253034] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.253034] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 583.253511] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.253511] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 583.253511] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2d8e87a1-a834-438b-a0ca-187f648f9236, please check neutron logs for more information. [ 583.253511] env[63355]: ERROR nova.compute.manager [ 583.253511] env[63355]: Traceback (most recent call last): [ 583.253511] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 583.253511] env[63355]: listener.cb(fileno) [ 583.253511] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.253511] env[63355]: result = function(*args, **kwargs) [ 583.253511] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.253511] env[63355]: return func(*args, **kwargs) [ 583.253511] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.253511] env[63355]: raise e [ 583.253511] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.253511] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 583.253511] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.253511] env[63355]: created_port_ids = self._update_ports_for_instance( [ 583.253511] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.253511] env[63355]: with excutils.save_and_reraise_exception(): [ 583.253511] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.253511] env[63355]: self.force_reraise() [ 583.253511] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.253511] env[63355]: raise self.value [ 583.253511] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.253511] env[63355]: updated_port = self._update_port( [ 583.253511] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.253511] env[63355]: _ensure_no_port_binding_failure(port) [ 583.253511] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.253511] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 583.256285] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 2d8e87a1-a834-438b-a0ca-187f648f9236, please check neutron logs for more information. [ 583.256285] env[63355]: Removing descriptor: 14 [ 583.259422] env[63355]: ERROR nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2d8e87a1-a834-438b-a0ca-187f648f9236, please check neutron logs for more information. [ 583.259422] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Traceback (most recent call last): [ 583.259422] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 583.259422] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] yield resources [ 583.259422] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.259422] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] self.driver.spawn(context, instance, image_meta, [ 583.259422] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 583.259422] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.259422] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.259422] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] vm_ref = self.build_virtual_machine(instance, [ 583.259422] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] for vif in network_info: [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] return self._sync_wrapper(fn, *args, **kwargs) [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] self.wait() [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] self[:] = self._gt.wait() [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] return self._exit_event.wait() [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.259917] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] result = hub.switch() [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] return self.greenlet.switch() [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] result = function(*args, **kwargs) [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] return func(*args, **kwargs) [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] raise e [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] nwinfo = self.network_api.allocate_for_instance( [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] created_port_ids = self._update_ports_for_instance( [ 583.260242] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] with excutils.save_and_reraise_exception(): [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] self.force_reraise() [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] raise self.value [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] updated_port = self._update_port( [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] _ensure_no_port_binding_failure(port) [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] raise exception.PortBindingFailed(port_id=port['id']) [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] nova.exception.PortBindingFailed: Binding failed for port 2d8e87a1-a834-438b-a0ca-187f648f9236, please check neutron logs for more information. [ 583.260588] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] [ 583.260977] env[63355]: INFO nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Terminating instance [ 583.262978] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Acquiring lock "refresh_cache-e86f2c82-d544-4d73-bcc4-f55edf56cab3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.262978] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Acquired lock "refresh_cache-e86f2c82-d544-4d73-bcc4-f55edf56cab3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.262978] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.313801] env[63355]: DEBUG nova.compute.utils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 583.315578] env[63355]: DEBUG nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 583.315685] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 583.431579] env[63355]: DEBUG nova.policy [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe6b27b9050140f095dee9054ddee275', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '47cc7494ec7048eabe549433109dfda6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 583.525318] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.819042] env[63355]: DEBUG nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.830728] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.946057] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac83d9a5-1b99-466b-af95-379bf8e28eb3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.954714] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1fe87f-139f-4e67-9491-e976fbe43288 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.989306] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0553e98-dd2e-4f86-9cc8-b24eaefc500d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.997402] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29bcfd7c-fd47-4ecf-a8ff-f4a193b87e0c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.013274] env[63355]: DEBUG nova.compute.provider_tree [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.264725] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.379989] env[63355]: DEBUG nova.compute.manager [req-879e908d-22d5-4183-b3cb-2d6b8e749c92 req-52424ba0-33ce-4813-be36-d6ead4264090 service nova] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Received event network-changed-2d8e87a1-a834-438b-a0ca-187f648f9236 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.379989] env[63355]: DEBUG nova.compute.manager [req-879e908d-22d5-4183-b3cb-2d6b8e749c92 req-52424ba0-33ce-4813-be36-d6ead4264090 service nova] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Refreshing instance network info cache due to event network-changed-2d8e87a1-a834-438b-a0ca-187f648f9236. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 584.379989] env[63355]: DEBUG oslo_concurrency.lockutils [req-879e908d-22d5-4183-b3cb-2d6b8e749c92 req-52424ba0-33ce-4813-be36-d6ead4264090 service nova] Acquiring lock "refresh_cache-e86f2c82-d544-4d73-bcc4-f55edf56cab3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.514232] env[63355]: DEBUG nova.scheduler.client.report [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.709140] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Successfully created port: d96025ae-f12d-493d-90c3-8d3fefd523e4 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 584.780420] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Releasing lock "refresh_cache-e86f2c82-d544-4d73-bcc4-f55edf56cab3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.781187] env[63355]: DEBUG nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 584.781588] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 584.786374] env[63355]: DEBUG oslo_concurrency.lockutils [req-879e908d-22d5-4183-b3cb-2d6b8e749c92 req-52424ba0-33ce-4813-be36-d6ead4264090 service nova] Acquired lock "refresh_cache-e86f2c82-d544-4d73-bcc4-f55edf56cab3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.786374] env[63355]: DEBUG nova.network.neutron [req-879e908d-22d5-4183-b3cb-2d6b8e749c92 req-52424ba0-33ce-4813-be36-d6ead4264090 service nova] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Refreshing network info cache for port 2d8e87a1-a834-438b-a0ca-187f648f9236 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 584.790223] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63439038-af34-44b6-b478-6781e3f95a34 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.805793] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193d9081-2ff3-4c8a-b6c5-9d69b12f505d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.835474] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e86f2c82-d544-4d73-bcc4-f55edf56cab3 could not be found. [ 584.836593] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.836593] env[63355]: INFO nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 584.836745] env[63355]: DEBUG oslo.service.loopingcall [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.837075] env[63355]: DEBUG nova.compute.manager [-] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.837075] env[63355]: DEBUG nova.network.neutron [-] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.845975] env[63355]: DEBUG nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 584.884174] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.885997] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.887138] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.887884] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.888386] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.889226] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.889590] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.893021] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.893021] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.893021] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.893021] env[63355]: DEBUG nova.virt.hardware [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.893021] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe76382-d048-4909-ba52-592425c3b5f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.898667] env[63355]: DEBUG nova.network.neutron [-] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.906437] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608249a4-98d6-4668-8f74-df672f72e821 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.915785] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "ce2f828a-82d0-4270-ac21-d967b2e0594b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.915999] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "ce2f828a-82d0-4270-ac21-d967b2e0594b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.021921] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.214s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.022529] env[63355]: DEBUG nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.025290] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.504s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.026726] env[63355]: INFO nova.compute.claims [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.364189] env[63355]: DEBUG nova.network.neutron [req-879e908d-22d5-4183-b3cb-2d6b8e749c92 req-52424ba0-33ce-4813-be36-d6ead4264090 service nova] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.402191] env[63355]: DEBUG nova.network.neutron [-] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.428656] env[63355]: DEBUG nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 585.530757] env[63355]: DEBUG nova.compute.utils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.534998] env[63355]: DEBUG nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.535187] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 585.673666] env[63355]: DEBUG nova.policy [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4fb9862ade6b4550a2c8ae07ef90402e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f7844ba45d044bbadd83ac3e0700b61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.908635] env[63355]: INFO nova.compute.manager [-] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Took 1.07 seconds to deallocate network for instance. [ 585.913421] env[63355]: DEBUG nova.compute.claims [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.913612] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.974964] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.003657] env[63355]: DEBUG nova.network.neutron [req-879e908d-22d5-4183-b3cb-2d6b8e749c92 req-52424ba0-33ce-4813-be36-d6ead4264090 service nova] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.036190] env[63355]: DEBUG nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.174643] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a615dc7-658e-4ac4-bc79-c6ec99c9f0f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.183695] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf80e35-03b7-44c0-9a10-cd6d5d38350c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.219455] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c084858-71c0-4d0f-a6ad-3b944ecfcc43 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.228279] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702969fb-7359-4c45-bcba-e88a9d413130 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.244286] env[63355]: DEBUG nova.compute.provider_tree [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.507950] env[63355]: DEBUG oslo_concurrency.lockutils [req-879e908d-22d5-4183-b3cb-2d6b8e749c92 req-52424ba0-33ce-4813-be36-d6ead4264090 service nova] Releasing lock "refresh_cache-e86f2c82-d544-4d73-bcc4-f55edf56cab3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.513300] env[63355]: ERROR nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6038aaf0-9377-44f3-adc1-bb4c25210d18, please check neutron logs for more information. [ 586.513300] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 586.513300] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.513300] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 586.513300] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.513300] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 586.513300] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.513300] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 586.513300] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.513300] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 586.513300] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.513300] env[63355]: ERROR nova.compute.manager raise self.value [ 586.513300] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.513300] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 586.513300] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.513300] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 586.514170] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.514170] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 586.514170] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6038aaf0-9377-44f3-adc1-bb4c25210d18, please check neutron logs for more information. [ 586.514170] env[63355]: ERROR nova.compute.manager [ 586.514170] env[63355]: Traceback (most recent call last): [ 586.514170] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 586.514170] env[63355]: listener.cb(fileno) [ 586.514170] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.514170] env[63355]: result = function(*args, **kwargs) [ 586.514170] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.514170] env[63355]: return func(*args, **kwargs) [ 586.514170] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.514170] env[63355]: raise e [ 586.514170] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.514170] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 586.514170] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.514170] env[63355]: created_port_ids = self._update_ports_for_instance( [ 586.514170] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.514170] env[63355]: with excutils.save_and_reraise_exception(): [ 586.514170] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.514170] env[63355]: self.force_reraise() [ 586.514170] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.514170] env[63355]: raise self.value [ 586.514170] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.514170] env[63355]: updated_port = self._update_port( [ 586.514170] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.514170] env[63355]: _ensure_no_port_binding_failure(port) [ 586.514170] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.514170] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 586.515144] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 6038aaf0-9377-44f3-adc1-bb4c25210d18, please check neutron logs for more information. [ 586.515144] env[63355]: Removing descriptor: 16 [ 586.515144] env[63355]: ERROR nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6038aaf0-9377-44f3-adc1-bb4c25210d18, please check neutron logs for more information. [ 586.515144] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Traceback (most recent call last): [ 586.515144] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 586.515144] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] yield resources [ 586.515144] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.515144] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] self.driver.spawn(context, instance, image_meta, [ 586.515144] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 586.515144] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.515144] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.515144] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] vm_ref = self.build_virtual_machine(instance, [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] for vif in network_info: [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] return self._sync_wrapper(fn, *args, **kwargs) [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] self.wait() [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] self[:] = self._gt.wait() [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] return self._exit_event.wait() [ 586.515446] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] result = hub.switch() [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] return self.greenlet.switch() [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] result = function(*args, **kwargs) [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] return func(*args, **kwargs) [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] raise e [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] nwinfo = self.network_api.allocate_for_instance( [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.516320] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] created_port_ids = self._update_ports_for_instance( [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] with excutils.save_and_reraise_exception(): [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] self.force_reraise() [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] raise self.value [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] updated_port = self._update_port( [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] _ensure_no_port_binding_failure(port) [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.517299] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] raise exception.PortBindingFailed(port_id=port['id']) [ 586.517817] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] nova.exception.PortBindingFailed: Binding failed for port 6038aaf0-9377-44f3-adc1-bb4c25210d18, please check neutron logs for more information. [ 586.517817] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] [ 586.517817] env[63355]: INFO nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Terminating instance [ 586.517817] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Acquiring lock "refresh_cache-35bd361f-fda1-4121-9f4c-61d3219f9202" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.517817] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Acquired lock "refresh_cache-35bd361f-fda1-4121-9f4c-61d3219f9202" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.517817] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 586.609508] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquiring lock "abe6726e-7d6a-46e0-9273-de3e681311eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.609740] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lock "abe6726e-7d6a-46e0-9273-de3e681311eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.748927] env[63355]: DEBUG nova.scheduler.client.report [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.948495] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Successfully created port: 16b85a19-bea8-4470-82d6-d4c17ade8cf1 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 587.055987] env[63355]: DEBUG nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.091768] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.092374] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.092678] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.092979] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.093276] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.093555] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.093875] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.094153] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.094467] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.094758] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.095057] env[63355]: DEBUG nova.virt.hardware [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.096230] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e551b33-67ba-4759-aa8c-8dea8634942b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.105029] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113f78ab-3819-44b0-9cbb-2405ebaf48d0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.122384] env[63355]: DEBUG nova.compute.manager [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.166374] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.259243] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.231s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.259243] env[63355]: DEBUG nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 587.260204] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 1.346s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.411637] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.517198] env[63355]: DEBUG nova.compute.manager [req-95f93fde-07c8-49b2-ad4f-1c1a013c0856 req-53cd7422-e161-4925-8c05-686d0c320a6f service nova] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Received event network-changed-6038aaf0-9377-44f3-adc1-bb4c25210d18 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.517728] env[63355]: DEBUG nova.compute.manager [req-95f93fde-07c8-49b2-ad4f-1c1a013c0856 req-53cd7422-e161-4925-8c05-686d0c320a6f service nova] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Refreshing instance network info cache due to event network-changed-6038aaf0-9377-44f3-adc1-bb4c25210d18. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 587.518149] env[63355]: DEBUG oslo_concurrency.lockutils [req-95f93fde-07c8-49b2-ad4f-1c1a013c0856 req-53cd7422-e161-4925-8c05-686d0c320a6f service nova] Acquiring lock "refresh_cache-35bd361f-fda1-4121-9f4c-61d3219f9202" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.537617] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Acquiring lock "8d2ea6a5-5c57-468f-be08-62f418b80e1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.537922] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Lock "8d2ea6a5-5c57-468f-be08-62f418b80e1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.649621] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.774130] env[63355]: DEBUG nova.compute.utils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 587.775071] env[63355]: DEBUG nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 587.775276] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 587.918668] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Releasing lock "refresh_cache-35bd361f-fda1-4121-9f4c-61d3219f9202" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.918668] env[63355]: DEBUG nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 587.918668] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 587.918668] env[63355]: DEBUG oslo_concurrency.lockutils [req-95f93fde-07c8-49b2-ad4f-1c1a013c0856 req-53cd7422-e161-4925-8c05-686d0c320a6f service nova] Acquired lock "refresh_cache-35bd361f-fda1-4121-9f4c-61d3219f9202" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.918668] env[63355]: DEBUG nova.network.neutron [req-95f93fde-07c8-49b2-ad4f-1c1a013c0856 req-53cd7422-e161-4925-8c05-686d0c320a6f service nova] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Refreshing network info cache for port 6038aaf0-9377-44f3-adc1-bb4c25210d18 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 587.919380] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-365b9019-f7be-4838-96ac-dd2b01eac4d9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.933871] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7902dedc-792f-43e2-94cf-fc1efbf12fad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.955180] env[63355]: DEBUG nova.policy [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dde65519973e474cb6d4acb7d37ed41a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f28f90c76b8345c5991b6b8bd54f0237', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 587.971337] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 35bd361f-fda1-4121-9f4c-61d3219f9202 could not be found. [ 587.971943] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 587.971943] env[63355]: INFO nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Took 0.05 seconds to destroy the instance on the hypervisor. [ 587.972245] env[63355]: DEBUG oslo.service.loopingcall [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.972994] env[63355]: DEBUG nova.compute.manager [-] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.973113] env[63355]: DEBUG nova.network.neutron [-] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.990881] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Acquiring lock "0fb6ba21-d45e-4809-8e68-d78a5892baeb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.990881] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Lock "0fb6ba21-d45e-4809-8e68-d78a5892baeb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.006612] env[63355]: DEBUG nova.network.neutron [-] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.027683] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7d186b-2668-49a9-a2bf-867abbcd36b1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.036233] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28e1952-6a09-4ae0-9038-7aba5f950ded {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.045944] env[63355]: DEBUG nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.076559] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157e6389-5bdb-40c3-84a4-6e50965899b8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.084833] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef1bbe0-fe3a-4ad4-b6c8-ef9015fc9166 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.106820] env[63355]: DEBUG nova.compute.provider_tree [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.172021] env[63355]: ERROR nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d96025ae-f12d-493d-90c3-8d3fefd523e4, please check neutron logs for more information. [ 588.172021] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 588.172021] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.172021] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 588.172021] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.172021] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 588.172021] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.172021] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 588.172021] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.172021] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 588.172021] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.172021] env[63355]: ERROR nova.compute.manager raise self.value [ 588.172021] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.172021] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 588.172021] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.172021] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 588.172675] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.172675] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 588.172675] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d96025ae-f12d-493d-90c3-8d3fefd523e4, please check neutron logs for more information. [ 588.172675] env[63355]: ERROR nova.compute.manager [ 588.172675] env[63355]: Traceback (most recent call last): [ 588.172675] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 588.172675] env[63355]: listener.cb(fileno) [ 588.172675] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.172675] env[63355]: result = function(*args, **kwargs) [ 588.172675] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.172675] env[63355]: return func(*args, **kwargs) [ 588.172675] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.172675] env[63355]: raise e [ 588.172675] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.172675] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 588.172675] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.172675] env[63355]: created_port_ids = self._update_ports_for_instance( [ 588.172675] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.172675] env[63355]: with excutils.save_and_reraise_exception(): [ 588.172675] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.172675] env[63355]: self.force_reraise() [ 588.172675] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.172675] env[63355]: raise self.value [ 588.172675] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.172675] env[63355]: updated_port = self._update_port( [ 588.172675] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.172675] env[63355]: _ensure_no_port_binding_failure(port) [ 588.172675] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.172675] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 588.173787] env[63355]: nova.exception.PortBindingFailed: Binding failed for port d96025ae-f12d-493d-90c3-8d3fefd523e4, please check neutron logs for more information. [ 588.173787] env[63355]: Removing descriptor: 14 [ 588.173787] env[63355]: ERROR nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d96025ae-f12d-493d-90c3-8d3fefd523e4, please check neutron logs for more information. [ 588.173787] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Traceback (most recent call last): [ 588.173787] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 588.173787] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] yield resources [ 588.173787] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.173787] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] self.driver.spawn(context, instance, image_meta, [ 588.173787] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 588.173787] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.173787] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.173787] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] vm_ref = self.build_virtual_machine(instance, [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] for vif in network_info: [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] return self._sync_wrapper(fn, *args, **kwargs) [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] self.wait() [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] self[:] = self._gt.wait() [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] return self._exit_event.wait() [ 588.174149] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] result = hub.switch() [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] return self.greenlet.switch() [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] result = function(*args, **kwargs) [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] return func(*args, **kwargs) [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] raise e [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] nwinfo = self.network_api.allocate_for_instance( [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.174492] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] created_port_ids = self._update_ports_for_instance( [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] with excutils.save_and_reraise_exception(): [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] self.force_reraise() [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] raise self.value [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] updated_port = self._update_port( [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] _ensure_no_port_binding_failure(port) [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.174884] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] raise exception.PortBindingFailed(port_id=port['id']) [ 588.175265] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] nova.exception.PortBindingFailed: Binding failed for port d96025ae-f12d-493d-90c3-8d3fefd523e4, please check neutron logs for more information. [ 588.175265] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] [ 588.175265] env[63355]: INFO nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Terminating instance [ 588.177958] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquiring lock "refresh_cache-b3d31162-5851-466f-b44c-d15e69c76bdd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.177958] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquired lock "refresh_cache-b3d31162-5851-466f-b44c-d15e69c76bdd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.177958] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.284335] env[63355]: DEBUG nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 588.494571] env[63355]: DEBUG nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.514428] env[63355]: DEBUG nova.network.neutron [-] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.598637] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.610074] env[63355]: DEBUG nova.scheduler.client.report [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.626966] env[63355]: DEBUG nova.network.neutron [req-95f93fde-07c8-49b2-ad4f-1c1a013c0856 req-53cd7422-e161-4925-8c05-686d0c320a6f service nova] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.634951] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Successfully created port: f270dd28-d082-4021-8f15-d71f9eec8b01 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.729984] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.965250] env[63355]: DEBUG nova.network.neutron [req-95f93fde-07c8-49b2-ad4f-1c1a013c0856 req-53cd7422-e161-4925-8c05-686d0c320a6f service nova] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.984202] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.018392] env[63355]: INFO nova.compute.manager [-] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Took 1.04 seconds to deallocate network for instance. [ 589.021849] env[63355]: DEBUG nova.compute.claims [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 589.022037] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.023981] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.116372] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.856s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.117199] env[63355]: ERROR nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2d8e87a1-a834-438b-a0ca-187f648f9236, please check neutron logs for more information. [ 589.117199] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Traceback (most recent call last): [ 589.117199] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.117199] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] self.driver.spawn(context, instance, image_meta, [ 589.117199] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 589.117199] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.117199] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.117199] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] vm_ref = self.build_virtual_machine(instance, [ 589.117199] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.117199] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.117199] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] for vif in network_info: [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] return self._sync_wrapper(fn, *args, **kwargs) [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] self.wait() [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] self[:] = self._gt.wait() [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] return self._exit_event.wait() [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] result = hub.switch() [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.117556] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] return self.greenlet.switch() [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] result = function(*args, **kwargs) [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] return func(*args, **kwargs) [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] raise e [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] nwinfo = self.network_api.allocate_for_instance( [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] created_port_ids = self._update_ports_for_instance( [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] with excutils.save_and_reraise_exception(): [ 589.117918] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] self.force_reraise() [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] raise self.value [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] updated_port = self._update_port( [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] _ensure_no_port_binding_failure(port) [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] raise exception.PortBindingFailed(port_id=port['id']) [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] nova.exception.PortBindingFailed: Binding failed for port 2d8e87a1-a834-438b-a0ca-187f648f9236, please check neutron logs for more information. [ 589.118278] env[63355]: ERROR nova.compute.manager [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] [ 589.123131] env[63355]: DEBUG nova.compute.utils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Binding failed for port 2d8e87a1-a834-438b-a0ca-187f648f9236, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 589.123183] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.149s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.124773] env[63355]: INFO nova.compute.claims [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.130040] env[63355]: DEBUG nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Build of instance e86f2c82-d544-4d73-bcc4-f55edf56cab3 was re-scheduled: Binding failed for port 2d8e87a1-a834-438b-a0ca-187f648f9236, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 589.130514] env[63355]: DEBUG nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 589.130762] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Acquiring lock "refresh_cache-e86f2c82-d544-4d73-bcc4-f55edf56cab3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.130908] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Acquired lock "refresh_cache-e86f2c82-d544-4d73-bcc4-f55edf56cab3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.131077] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.293747] env[63355]: DEBUG nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 589.329198] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.329198] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.329953] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.330494] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.330894] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.331205] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.331664] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.332108] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.332867] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.332867] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.333045] env[63355]: DEBUG nova.virt.hardware [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.334117] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b277fc-5554-47b8-b1c7-c2c740d2c1b1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.344235] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8697e1b-f458-4e25-8df8-7a1e1640e02c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.469040] env[63355]: DEBUG oslo_concurrency.lockutils [req-95f93fde-07c8-49b2-ad4f-1c1a013c0856 req-53cd7422-e161-4925-8c05-686d0c320a6f service nova] Releasing lock "refresh_cache-35bd361f-fda1-4121-9f4c-61d3219f9202" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.492335] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Releasing lock "refresh_cache-b3d31162-5851-466f-b44c-d15e69c76bdd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.492335] env[63355]: DEBUG nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.492335] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.492335] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91560f51-9b53-477c-9104-86de3b0ca0d1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.502606] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38989388-9b5b-4332-b994-11a4e869f8bd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.527942] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b3d31162-5851-466f-b44c-d15e69c76bdd could not be found. [ 589.528321] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 589.528559] env[63355]: INFO nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 589.528937] env[63355]: DEBUG oslo.service.loopingcall [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.529229] env[63355]: DEBUG nova.compute.manager [-] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.529389] env[63355]: DEBUG nova.network.neutron [-] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.611563] env[63355]: DEBUG nova.network.neutron [-] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.795779] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.053203] env[63355]: ERROR nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 16b85a19-bea8-4470-82d6-d4c17ade8cf1, please check neutron logs for more information. [ 590.053203] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 590.053203] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.053203] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 590.053203] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.053203] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 590.053203] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.053203] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 590.053203] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.053203] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 590.053203] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.053203] env[63355]: ERROR nova.compute.manager raise self.value [ 590.053203] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.053203] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 590.053203] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.053203] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 590.053960] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.053960] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 590.053960] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 16b85a19-bea8-4470-82d6-d4c17ade8cf1, please check neutron logs for more information. [ 590.053960] env[63355]: ERROR nova.compute.manager [ 590.053960] env[63355]: Traceback (most recent call last): [ 590.053960] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 590.053960] env[63355]: listener.cb(fileno) [ 590.053960] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.053960] env[63355]: result = function(*args, **kwargs) [ 590.053960] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.053960] env[63355]: return func(*args, **kwargs) [ 590.053960] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.053960] env[63355]: raise e [ 590.053960] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.053960] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 590.053960] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.053960] env[63355]: created_port_ids = self._update_ports_for_instance( [ 590.053960] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.053960] env[63355]: with excutils.save_and_reraise_exception(): [ 590.053960] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.053960] env[63355]: self.force_reraise() [ 590.053960] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.053960] env[63355]: raise self.value [ 590.053960] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.053960] env[63355]: updated_port = self._update_port( [ 590.053960] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.053960] env[63355]: _ensure_no_port_binding_failure(port) [ 590.053960] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.053960] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 590.054754] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 16b85a19-bea8-4470-82d6-d4c17ade8cf1, please check neutron logs for more information. [ 590.054754] env[63355]: Removing descriptor: 18 [ 590.054754] env[63355]: ERROR nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 16b85a19-bea8-4470-82d6-d4c17ade8cf1, please check neutron logs for more information. [ 590.054754] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Traceback (most recent call last): [ 590.054754] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 590.054754] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] yield resources [ 590.054754] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 590.054754] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] self.driver.spawn(context, instance, image_meta, [ 590.054754] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 590.054754] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.054754] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.054754] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] vm_ref = self.build_virtual_machine(instance, [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] for vif in network_info: [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] return self._sync_wrapper(fn, *args, **kwargs) [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] self.wait() [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] self[:] = self._gt.wait() [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] return self._exit_event.wait() [ 590.055145] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] result = hub.switch() [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] return self.greenlet.switch() [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] result = function(*args, **kwargs) [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] return func(*args, **kwargs) [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] raise e [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] nwinfo = self.network_api.allocate_for_instance( [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.055527] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] created_port_ids = self._update_ports_for_instance( [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] with excutils.save_and_reraise_exception(): [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] self.force_reraise() [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] raise self.value [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] updated_port = self._update_port( [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] _ensure_no_port_binding_failure(port) [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.055843] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] raise exception.PortBindingFailed(port_id=port['id']) [ 590.056144] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] nova.exception.PortBindingFailed: Binding failed for port 16b85a19-bea8-4470-82d6-d4c17ade8cf1, please check neutron logs for more information. [ 590.056144] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] [ 590.056144] env[63355]: INFO nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Terminating instance [ 590.058716] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Acquiring lock "refresh_cache-d08c461b-c2ba-4484-a0aa-746d726868a0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.058716] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Acquired lock "refresh_cache-d08c461b-c2ba-4484-a0aa-746d726868a0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.058716] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 590.113591] env[63355]: DEBUG nova.network.neutron [-] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.128701] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.171354] env[63355]: DEBUG nova.compute.manager [req-b1bfb1b4-08b3-4dc8-bd4e-3b3f36e19ab0 req-956716d6-56bd-4326-a46a-1cbaf24f5e01 service nova] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Received event network-vif-deleted-6038aaf0-9377-44f3-adc1-bb4c25210d18 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.427453] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d87aac-b8ae-414a-a26b-62a63673d64f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.436605] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0e976b-4310-4ccb-a434-a581f6bf8733 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.473353] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8589a4-c488-4e61-96ff-f9b248d696c9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.480454] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d858a0-cfc4-4fa9-8957-3ae3dd513642 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.495245] env[63355]: DEBUG nova.compute.provider_tree [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.605726] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.620545] env[63355]: INFO nova.compute.manager [-] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Took 1.09 seconds to deallocate network for instance. [ 590.626196] env[63355]: DEBUG nova.compute.claims [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.626196] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.632964] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Releasing lock "refresh_cache-e86f2c82-d544-4d73-bcc4-f55edf56cab3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.632964] env[63355]: DEBUG nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 590.632964] env[63355]: DEBUG nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.632964] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.639137] env[63355]: DEBUG nova.compute.manager [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Received event network-changed-d96025ae-f12d-493d-90c3-8d3fefd523e4 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.639311] env[63355]: DEBUG nova.compute.manager [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Refreshing instance network info cache due to event network-changed-d96025ae-f12d-493d-90c3-8d3fefd523e4. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 590.639507] env[63355]: DEBUG oslo_concurrency.lockutils [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] Acquiring lock "refresh_cache-b3d31162-5851-466f-b44c-d15e69c76bdd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.639699] env[63355]: DEBUG oslo_concurrency.lockutils [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] Acquired lock "refresh_cache-b3d31162-5851-466f-b44c-d15e69c76bdd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.641543] env[63355]: DEBUG nova.network.neutron [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Refreshing network info cache for port d96025ae-f12d-493d-90c3-8d3fefd523e4 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 590.719345] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.837739] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.001756] env[63355]: DEBUG nova.scheduler.client.report [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.177685] env[63355]: DEBUG nova.network.neutron [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.221627] env[63355]: DEBUG nova.network.neutron [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.344018] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Releasing lock "refresh_cache-d08c461b-c2ba-4484-a0aa-746d726868a0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.344018] env[63355]: DEBUG nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 591.344018] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 591.344018] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b616f51a-85cb-4ff4-b6b3-b85021338ee6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.354531] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571483ac-2802-449b-9353-b9cb177339c4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.379816] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d08c461b-c2ba-4484-a0aa-746d726868a0 could not be found. [ 591.379816] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 591.379816] env[63355]: INFO nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 591.379816] env[63355]: DEBUG oslo.service.loopingcall [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.379816] env[63355]: DEBUG nova.compute.manager [-] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.380055] env[63355]: DEBUG nova.network.neutron [-] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 591.401792] env[63355]: DEBUG nova.network.neutron [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.429468] env[63355]: DEBUG nova.network.neutron [-] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.430672] env[63355]: ERROR nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f270dd28-d082-4021-8f15-d71f9eec8b01, please check neutron logs for more information. [ 591.430672] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 591.430672] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.430672] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 591.430672] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.430672] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 591.430672] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.430672] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 591.430672] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.430672] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 591.430672] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.430672] env[63355]: ERROR nova.compute.manager raise self.value [ 591.430672] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.430672] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 591.430672] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.430672] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 591.431138] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.431138] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 591.431138] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f270dd28-d082-4021-8f15-d71f9eec8b01, please check neutron logs for more information. [ 591.431138] env[63355]: ERROR nova.compute.manager [ 591.431138] env[63355]: Traceback (most recent call last): [ 591.431138] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 591.431138] env[63355]: listener.cb(fileno) [ 591.431138] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.431138] env[63355]: result = function(*args, **kwargs) [ 591.431138] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.431138] env[63355]: return func(*args, **kwargs) [ 591.431138] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.431138] env[63355]: raise e [ 591.431138] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.431138] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 591.431138] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.431138] env[63355]: created_port_ids = self._update_ports_for_instance( [ 591.431138] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.431138] env[63355]: with excutils.save_and_reraise_exception(): [ 591.431138] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.431138] env[63355]: self.force_reraise() [ 591.431138] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.431138] env[63355]: raise self.value [ 591.431138] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.431138] env[63355]: updated_port = self._update_port( [ 591.431138] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.431138] env[63355]: _ensure_no_port_binding_failure(port) [ 591.431138] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.431138] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 591.431839] env[63355]: nova.exception.PortBindingFailed: Binding failed for port f270dd28-d082-4021-8f15-d71f9eec8b01, please check neutron logs for more information. [ 591.431839] env[63355]: Removing descriptor: 16 [ 591.435373] env[63355]: ERROR nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f270dd28-d082-4021-8f15-d71f9eec8b01, please check neutron logs for more information. [ 591.435373] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Traceback (most recent call last): [ 591.435373] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 591.435373] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] yield resources [ 591.435373] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.435373] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] self.driver.spawn(context, instance, image_meta, [ 591.435373] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 591.435373] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.435373] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.435373] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] vm_ref = self.build_virtual_machine(instance, [ 591.435373] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] for vif in network_info: [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] return self._sync_wrapper(fn, *args, **kwargs) [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] self.wait() [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] self[:] = self._gt.wait() [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] return self._exit_event.wait() [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.435801] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] result = hub.switch() [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] return self.greenlet.switch() [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] result = function(*args, **kwargs) [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] return func(*args, **kwargs) [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] raise e [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] nwinfo = self.network_api.allocate_for_instance( [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] created_port_ids = self._update_ports_for_instance( [ 591.436170] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] with excutils.save_and_reraise_exception(): [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] self.force_reraise() [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] raise self.value [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] updated_port = self._update_port( [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] _ensure_no_port_binding_failure(port) [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] raise exception.PortBindingFailed(port_id=port['id']) [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] nova.exception.PortBindingFailed: Binding failed for port f270dd28-d082-4021-8f15-d71f9eec8b01, please check neutron logs for more information. [ 591.436588] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] [ 591.437071] env[63355]: INFO nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Terminating instance [ 591.437071] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquiring lock "refresh_cache-059dc72c-dc91-4744-8efa-886aef99d5c5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.437288] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquired lock "refresh_cache-059dc72c-dc91-4744-8efa-886aef99d5c5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.438605] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.506156] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.506611] env[63355]: DEBUG nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 591.509161] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.860s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.511011] env[63355]: INFO nova.compute.claims [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 591.726553] env[63355]: INFO nova.compute.manager [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] [instance: e86f2c82-d544-4d73-bcc4-f55edf56cab3] Took 1.09 seconds to deallocate network for instance. [ 591.906518] env[63355]: DEBUG oslo_concurrency.lockutils [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] Releasing lock "refresh_cache-b3d31162-5851-466f-b44c-d15e69c76bdd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.906793] env[63355]: DEBUG nova.compute.manager [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Received event network-vif-deleted-d96025ae-f12d-493d-90c3-8d3fefd523e4 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.906977] env[63355]: DEBUG nova.compute.manager [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Received event network-changed-16b85a19-bea8-4470-82d6-d4c17ade8cf1 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.907155] env[63355]: DEBUG nova.compute.manager [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Refreshing instance network info cache due to event network-changed-16b85a19-bea8-4470-82d6-d4c17ade8cf1. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 591.907376] env[63355]: DEBUG oslo_concurrency.lockutils [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] Acquiring lock "refresh_cache-d08c461b-c2ba-4484-a0aa-746d726868a0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.907516] env[63355]: DEBUG oslo_concurrency.lockutils [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] Acquired lock "refresh_cache-d08c461b-c2ba-4484-a0aa-746d726868a0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.907674] env[63355]: DEBUG nova.network.neutron [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Refreshing network info cache for port 16b85a19-bea8-4470-82d6-d4c17ade8cf1 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 591.934376] env[63355]: DEBUG nova.network.neutron [-] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.968163] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.016750] env[63355]: DEBUG nova.compute.utils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 592.020617] env[63355]: DEBUG nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 592.020789] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 592.034356] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.164925] env[63355]: DEBUG nova.policy [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7123da3ee74b497395f1681e460aaaca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b6963b8e16b4986a4545914b75a38ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 592.440135] env[63355]: INFO nova.compute.manager [-] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Took 1.06 seconds to deallocate network for instance. [ 592.441553] env[63355]: DEBUG nova.compute.claims [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 592.441899] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.504385] env[63355]: DEBUG nova.network.neutron [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.526727] env[63355]: DEBUG nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 592.538250] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Releasing lock "refresh_cache-059dc72c-dc91-4744-8efa-886aef99d5c5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.538250] env[63355]: DEBUG nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.538250] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 592.538250] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f251a2a-47dc-486f-be53-cc2e6b218a7f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.552022] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd1d408-d1e6-4dda-a6e4-fffd096653d3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.587657] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 059dc72c-dc91-4744-8efa-886aef99d5c5 could not be found. [ 592.588652] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 592.588652] env[63355]: INFO nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Took 0.05 seconds to destroy the instance on the hypervisor. [ 592.588652] env[63355]: DEBUG oslo.service.loopingcall [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.588652] env[63355]: DEBUG nova.compute.manager [-] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.588652] env[63355]: DEBUG nova.network.neutron [-] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.687260] env[63355]: DEBUG nova.network.neutron [-] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.709300] env[63355]: DEBUG nova.compute.manager [req-7b9ce8fe-26b0-42f3-8ed7-d5d5942fdcd3 req-678d15c5-0c84-4ccd-9a12-1e8da7e97268 service nova] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Received event network-changed-f270dd28-d082-4021-8f15-d71f9eec8b01 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.711017] env[63355]: DEBUG nova.compute.manager [req-7b9ce8fe-26b0-42f3-8ed7-d5d5942fdcd3 req-678d15c5-0c84-4ccd-9a12-1e8da7e97268 service nova] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Refreshing instance network info cache due to event network-changed-f270dd28-d082-4021-8f15-d71f9eec8b01. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 592.711017] env[63355]: DEBUG oslo_concurrency.lockutils [req-7b9ce8fe-26b0-42f3-8ed7-d5d5942fdcd3 req-678d15c5-0c84-4ccd-9a12-1e8da7e97268 service nova] Acquiring lock "refresh_cache-059dc72c-dc91-4744-8efa-886aef99d5c5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.711017] env[63355]: DEBUG oslo_concurrency.lockutils [req-7b9ce8fe-26b0-42f3-8ed7-d5d5942fdcd3 req-678d15c5-0c84-4ccd-9a12-1e8da7e97268 service nova] Acquired lock "refresh_cache-059dc72c-dc91-4744-8efa-886aef99d5c5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.711017] env[63355]: DEBUG nova.network.neutron [req-7b9ce8fe-26b0-42f3-8ed7-d5d5942fdcd3 req-678d15c5-0c84-4ccd-9a12-1e8da7e97268 service nova] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Refreshing network info cache for port f270dd28-d082-4021-8f15-d71f9eec8b01 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 592.745780] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339241a6-d525-447c-8158-b0297c553652 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.760021] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1970e3d-f870-4e45-ba1e-c74dd3cb0b23 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.808117] env[63355]: INFO nova.scheduler.client.report [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Deleted allocations for instance e86f2c82-d544-4d73-bcc4-f55edf56cab3 [ 592.813208] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b918a69-ffce-4e57-8b2d-67e08c8be697 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.821902] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f21e44-168e-43d1-821f-4c934e632d9a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.845536] env[63355]: DEBUG nova.compute.provider_tree [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.972480] env[63355]: DEBUG nova.network.neutron [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.190132] env[63355]: DEBUG nova.network.neutron [-] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.260353] env[63355]: DEBUG nova.network.neutron [req-7b9ce8fe-26b0-42f3-8ed7-d5d5942fdcd3 req-678d15c5-0c84-4ccd-9a12-1e8da7e97268 service nova] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.279815] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Successfully created port: f0619064-8c90-4914-9df4-a1e6286d3d89 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.319950] env[63355]: DEBUG oslo_concurrency.lockutils [None req-08bff7b8-e394-4f2c-94bc-0d18669eff9a tempest-ImagesOneServerNegativeTestJSON-967092576 tempest-ImagesOneServerNegativeTestJSON-967092576-project-member] Lock "e86f2c82-d544-4d73-bcc4-f55edf56cab3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.929s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.333519] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Acquiring lock "5c0ba7c6-9946-4765-be4b-d48d3823a68d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.333905] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Lock "5c0ba7c6-9946-4765-be4b-d48d3823a68d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.350067] env[63355]: DEBUG nova.scheduler.client.report [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.477892] env[63355]: DEBUG oslo_concurrency.lockutils [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] Releasing lock "refresh_cache-d08c461b-c2ba-4484-a0aa-746d726868a0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.477892] env[63355]: DEBUG nova.compute.manager [req-208ff916-2f20-43fd-94d6-4691cf8781f4 req-4be6a912-a494-465b-a8eb-2e50c034f958 service nova] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Received event network-vif-deleted-16b85a19-bea8-4470-82d6-d4c17ade8cf1 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.488450] env[63355]: DEBUG nova.network.neutron [req-7b9ce8fe-26b0-42f3-8ed7-d5d5942fdcd3 req-678d15c5-0c84-4ccd-9a12-1e8da7e97268 service nova] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.538279] env[63355]: DEBUG nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 593.573994] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 593.574277] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 593.574434] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 593.574610] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 593.574768] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 593.574896] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 593.575113] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 593.575270] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 593.575524] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 593.575585] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 593.575751] env[63355]: DEBUG nova.virt.hardware [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 593.576634] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59748f73-8a89-4650-838e-37af4ee949e2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.585159] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c452dc2-22fe-4353-9b1a-57da122a3d34 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.694994] env[63355]: INFO nova.compute.manager [-] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Took 1.11 seconds to deallocate network for instance. [ 593.697807] env[63355]: DEBUG nova.compute.claims [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 593.697905] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.842018] env[63355]: DEBUG nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.854645] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.855178] env[63355]: DEBUG nova.compute.manager [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 593.860145] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.261s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.861169] env[63355]: INFO nova.compute.claims [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.991774] env[63355]: DEBUG oslo_concurrency.lockutils [req-7b9ce8fe-26b0-42f3-8ed7-d5d5942fdcd3 req-678d15c5-0c84-4ccd-9a12-1e8da7e97268 service nova] Releasing lock "refresh_cache-059dc72c-dc91-4744-8efa-886aef99d5c5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.993109] env[63355]: DEBUG nova.compute.manager [req-7b9ce8fe-26b0-42f3-8ed7-d5d5942fdcd3 req-678d15c5-0c84-4ccd-9a12-1e8da7e97268 service nova] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Received event network-vif-deleted-f270dd28-d082-4021-8f15-d71f9eec8b01 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.365731] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.366925] env[63355]: DEBUG nova.compute.utils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 594.370364] env[63355]: DEBUG nova.compute.manager [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Not allocating networking since 'none' was specified. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 594.634243] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 594.634744] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 594.634744] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 594.635518] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Rebuilding the list of instances to heal {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 594.744853] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Acquiring lock "1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.745149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Lock "1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.874081] env[63355]: DEBUG nova.compute.manager [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 595.057014] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310c52bb-7764-456b-a516-d0c3df913653 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.066107] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdbcf98-370b-4dc3-b2d9-61bcef055372 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.099152] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b38224-9b61-4306-b82e-fa4a211812f7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.107025] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc70e8bf-4c06-40f8-8a55-a33e3812d9d9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.123063] env[63355]: DEBUG nova.compute.provider_tree [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.139901] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 595.140408] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 595.140632] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 595.140765] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 595.140889] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 595.141018] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 595.141249] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 595.141249] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Didn't find any instances for network info cache update. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 595.142111] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 595.142358] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 595.142549] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 595.142781] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 595.142965] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 595.143425] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._sync_power_states {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 595.248349] env[63355]: DEBUG nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.627058] env[63355]: DEBUG nova.scheduler.client.report [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.649209] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Getting list of instances from cluster (obj){ [ 595.649209] env[63355]: value = "domain-c8" [ 595.649209] env[63355]: _type = "ClusterComputeResource" [ 595.649209] env[63355]: } {{(pid=63355) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 595.649868] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152db898-8f72-4e76-ace9-5376187ecf5d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.667920] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Got total of 0 instances {{(pid=63355) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 595.667920] env[63355]: WARNING nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] While synchronizing instance power states, found 7 instances in the database and 0 instances on the hypervisor. [ 595.667920] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid 35bd361f-fda1-4121-9f4c-61d3219f9202 {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 595.667920] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid b3d31162-5851-466f-b44c-d15e69c76bdd {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 595.667920] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid d08c461b-c2ba-4484-a0aa-746d726868a0 {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 595.667920] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid 059dc72c-dc91-4744-8efa-886aef99d5c5 {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 595.667920] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid ce2f828a-82d0-4270-ac21-d967b2e0594b {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 595.668149] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid abe6726e-7d6a-46e0-9273-de3e681311eb {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 595.668149] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid 8d2ea6a5-5c57-468f-be08-62f418b80e1e {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 595.668149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "35bd361f-fda1-4121-9f4c-61d3219f9202" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.668149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "b3d31162-5851-466f-b44c-d15e69c76bdd" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.668149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "d08c461b-c2ba-4484-a0aa-746d726868a0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.668295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "059dc72c-dc91-4744-8efa-886aef99d5c5" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.668295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "ce2f828a-82d0-4270-ac21-d967b2e0594b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.668295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "abe6726e-7d6a-46e0-9273-de3e681311eb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.668295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "8d2ea6a5-5c57-468f-be08-62f418b80e1e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.668295] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 595.668433] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 595.668433] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 595.819040] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.886910] env[63355]: DEBUG nova.compute.manager [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 595.942044] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 595.942044] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 595.942229] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.942809] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 595.942809] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.942809] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 595.943068] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 595.943180] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 595.943438] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 595.943527] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 595.943683] env[63355]: DEBUG nova.virt.hardware [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 595.944627] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef276fd4-b42a-4e6c-86c9-aae1340d7e58 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.954501] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce8900b-b8ec-48c6-a725-d52d263bc92a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.972912] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 595.982098] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 595.982384] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7dc3e8bd-e65a-4051-a592-9f31571e93a4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.997773] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Created folder: OpenStack in parent group-v4. [ 595.998272] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Creating folder: Project (f0ec4c3ae53b4663a110377ac0e4fcc1). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 595.999077] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a33faa4-a83f-4c5f-a47f-90ac96af4245 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.009269] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Created folder: Project (f0ec4c3ae53b4663a110377ac0e4fcc1) in parent group-v287607. [ 596.009462] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Creating folder: Instances. Parent ref: group-v287608. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 596.009685] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aac2f948-f0bb-4d4e-9062-599be070e073 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.019043] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Created folder: Instances in parent group-v287608. [ 596.019043] env[63355]: DEBUG oslo.service.loopingcall [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.019043] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 596.019336] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1da2a76b-52db-4062-b144-b204a24c8e7a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.036483] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 596.036483] env[63355]: value = "task-1349463" [ 596.036483] env[63355]: _type = "Task" [ 596.036483] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.044852] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349463, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.135504] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.136693] env[63355]: DEBUG nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 596.140785] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.118s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.169113] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.524560] env[63355]: ERROR nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f0619064-8c90-4914-9df4-a1e6286d3d89, please check neutron logs for more information. [ 596.524560] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.524560] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.524560] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.524560] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.524560] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.524560] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.524560] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.524560] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.524560] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 596.524560] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.524560] env[63355]: ERROR nova.compute.manager raise self.value [ 596.524560] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.524560] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.524560] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.524560] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.525037] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.525037] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.525037] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f0619064-8c90-4914-9df4-a1e6286d3d89, please check neutron logs for more information. [ 596.525037] env[63355]: ERROR nova.compute.manager [ 596.525037] env[63355]: Traceback (most recent call last): [ 596.525037] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.525037] env[63355]: listener.cb(fileno) [ 596.525037] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.525037] env[63355]: result = function(*args, **kwargs) [ 596.525037] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.525037] env[63355]: return func(*args, **kwargs) [ 596.525037] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.525037] env[63355]: raise e [ 596.525037] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.525037] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 596.525037] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.525037] env[63355]: created_port_ids = self._update_ports_for_instance( [ 596.525037] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.525037] env[63355]: with excutils.save_and_reraise_exception(): [ 596.525037] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.525037] env[63355]: self.force_reraise() [ 596.525037] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.525037] env[63355]: raise self.value [ 596.525037] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.525037] env[63355]: updated_port = self._update_port( [ 596.525037] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.525037] env[63355]: _ensure_no_port_binding_failure(port) [ 596.525037] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.525037] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.525781] env[63355]: nova.exception.PortBindingFailed: Binding failed for port f0619064-8c90-4914-9df4-a1e6286d3d89, please check neutron logs for more information. [ 596.525781] env[63355]: Removing descriptor: 18 [ 596.525781] env[63355]: ERROR nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f0619064-8c90-4914-9df4-a1e6286d3d89, please check neutron logs for more information. [ 596.525781] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Traceback (most recent call last): [ 596.525781] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 596.525781] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] yield resources [ 596.525781] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.525781] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] self.driver.spawn(context, instance, image_meta, [ 596.525781] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 596.525781] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.525781] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.525781] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] vm_ref = self.build_virtual_machine(instance, [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] for vif in network_info: [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] return self._sync_wrapper(fn, *args, **kwargs) [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] self.wait() [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] self[:] = self._gt.wait() [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] return self._exit_event.wait() [ 596.526114] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] result = hub.switch() [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] return self.greenlet.switch() [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] result = function(*args, **kwargs) [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] return func(*args, **kwargs) [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] raise e [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] nwinfo = self.network_api.allocate_for_instance( [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.526448] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] created_port_ids = self._update_ports_for_instance( [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] with excutils.save_and_reraise_exception(): [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] self.force_reraise() [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] raise self.value [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] updated_port = self._update_port( [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] _ensure_no_port_binding_failure(port) [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.526789] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] raise exception.PortBindingFailed(port_id=port['id']) [ 596.527178] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] nova.exception.PortBindingFailed: Binding failed for port f0619064-8c90-4914-9df4-a1e6286d3d89, please check neutron logs for more information. [ 596.527178] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] [ 596.527178] env[63355]: INFO nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Terminating instance [ 596.527178] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "refresh_cache-ce2f828a-82d0-4270-ac21-d967b2e0594b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.527178] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "refresh_cache-ce2f828a-82d0-4270-ac21-d967b2e0594b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.527178] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.553605] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349463, 'name': CreateVM_Task, 'duration_secs': 0.366645} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.553605] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 596.554517] env[63355]: DEBUG oslo_vmware.service [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac71d992-8f08-4f5e-99a1-f09dc26b927d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.561228] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.561389] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.562687] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 596.562687] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74a0f794-9c08-4bf6-bdba-46a466693902 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.568242] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 596.568242] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5204bf12-77c6-323f-9787-e2a8d7707335" [ 596.568242] env[63355]: _type = "Task" [ 596.568242] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.583188] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5204bf12-77c6-323f-9787-e2a8d7707335, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.645636] env[63355]: DEBUG nova.compute.utils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.646924] env[63355]: DEBUG nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.651141] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 596.660461] env[63355]: DEBUG nova.compute.manager [req-45135d48-7609-42aa-b81d-173fbb2d9abe req-9b75763e-f912-4ac4-bcfb-ad3af9f1a104 service nova] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Received event network-changed-f0619064-8c90-4914-9df4-a1e6286d3d89 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.660563] env[63355]: DEBUG nova.compute.manager [req-45135d48-7609-42aa-b81d-173fbb2d9abe req-9b75763e-f912-4ac4-bcfb-ad3af9f1a104 service nova] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Refreshing instance network info cache due to event network-changed-f0619064-8c90-4914-9df4-a1e6286d3d89. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 596.661451] env[63355]: DEBUG oslo_concurrency.lockutils [req-45135d48-7609-42aa-b81d-173fbb2d9abe req-9b75763e-f912-4ac4-bcfb-ad3af9f1a104 service nova] Acquiring lock "refresh_cache-ce2f828a-82d0-4270-ac21-d967b2e0594b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.749630] env[63355]: DEBUG nova.policy [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ca1cfded32146b8956ea03b7eeeb7c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '260232c1a95c4248aa86870b249159c8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.838759] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9049de30-5bff-41be-88a2-3deabe6e95b1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.846917] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83256efe-92a9-4fb0-bc82-00f50aec5511 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.885318] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229f1524-8201-412c-ba65-d7daf8884a6b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.893095] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030fdc2e-fe2a-414e-b283-10d6d85ed2e2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.907453] env[63355]: DEBUG nova.compute.provider_tree [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.051042] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.083186] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.083453] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 597.083700] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.083835] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.084371] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 597.085519] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1aa29fe1-b47a-4829-85f2-4b784074ccc5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.095423] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 597.095608] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 597.096407] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0bde60-cad8-45c5-9b15-f135101ee20c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.103965] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f5af384-2e7a-4f6e-a135-f00240692f28 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.110122] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 597.110122] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525d2dca-e521-8731-4c16-2a00d56acef8" [ 597.110122] env[63355]: _type = "Task" [ 597.110122] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.120581] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525d2dca-e521-8731-4c16-2a00d56acef8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.162417] env[63355]: DEBUG nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 597.203864] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.254768] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Successfully created port: 29895cad-8791-4e5d-8afc-ea7cd1027834 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.412040] env[63355]: DEBUG nova.scheduler.client.report [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.627086] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Preparing fetch location {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 597.627421] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Creating directory with path [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 597.627630] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-176602f2-2034-4d4c-a1da-cfd0277eca9e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.651689] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Created directory with path [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 597.651921] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Fetch image to [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 597.651921] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Downloading image file data 84ca0bb3-9916-4cea-9399-75af77b0558e to [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk on the data store datastore2 {{(pid=63355) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 597.652771] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7d60d6-5629-4e15-b305-82dabe9e22c8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.669015] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7054a28f-d0c1-4cc1-8486-43957de710e7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.689186] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f7bef1a-9ed0-4c79-9416-6736f5d4df5e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.724983] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "refresh_cache-ce2f828a-82d0-4270-ac21-d967b2e0594b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.725421] env[63355]: DEBUG nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 597.725596] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 597.726148] env[63355]: DEBUG oslo_concurrency.lockutils [req-45135d48-7609-42aa-b81d-173fbb2d9abe req-9b75763e-f912-4ac4-bcfb-ad3af9f1a104 service nova] Acquired lock "refresh_cache-ce2f828a-82d0-4270-ac21-d967b2e0594b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.726321] env[63355]: DEBUG nova.network.neutron [req-45135d48-7609-42aa-b81d-173fbb2d9abe req-9b75763e-f912-4ac4-bcfb-ad3af9f1a104 service nova] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Refreshing network info cache for port f0619064-8c90-4914-9df4-a1e6286d3d89 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 597.727516] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5825d37f-40c6-4968-b954-15ed5e2639f6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.732512] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdd55db-b851-46a3-b6b7-c9ffcbd8cffc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.745068] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a432746-b64c-44b9-8019-c601a2b560df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.759757] env[63355]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-17d3a639-47c2-404c-9f65-1e24f9385ff7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.768334] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ce2f828a-82d0-4270-ac21-d967b2e0594b could not be found. [ 597.769109] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 597.769379] env[63355]: INFO nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 597.770466] env[63355]: DEBUG oslo.service.loopingcall [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.770727] env[63355]: DEBUG nova.compute.manager [-] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 597.770818] env[63355]: DEBUG nova.network.neutron [-] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 597.813274] env[63355]: DEBUG nova.network.neutron [-] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.852396] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Downloading image file data 84ca0bb3-9916-4cea-9399-75af77b0558e to the data store datastore2 {{(pid=63355) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 597.919597] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.779s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.920420] env[63355]: ERROR nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6038aaf0-9377-44f3-adc1-bb4c25210d18, please check neutron logs for more information. [ 597.920420] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Traceback (most recent call last): [ 597.920420] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.920420] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] self.driver.spawn(context, instance, image_meta, [ 597.920420] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.920420] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.920420] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.920420] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] vm_ref = self.build_virtual_machine(instance, [ 597.920420] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.920420] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.920420] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] for vif in network_info: [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] return self._sync_wrapper(fn, *args, **kwargs) [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] self.wait() [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] self[:] = self._gt.wait() [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] return self._exit_event.wait() [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] result = hub.switch() [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.920781] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] return self.greenlet.switch() [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] result = function(*args, **kwargs) [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] return func(*args, **kwargs) [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] raise e [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] nwinfo = self.network_api.allocate_for_instance( [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] created_port_ids = self._update_ports_for_instance( [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] with excutils.save_and_reraise_exception(): [ 597.921174] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] self.force_reraise() [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] raise self.value [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] updated_port = self._update_port( [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] _ensure_no_port_binding_failure(port) [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] raise exception.PortBindingFailed(port_id=port['id']) [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] nova.exception.PortBindingFailed: Binding failed for port 6038aaf0-9377-44f3-adc1-bb4c25210d18, please check neutron logs for more information. [ 597.921527] env[63355]: ERROR nova.compute.manager [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] [ 597.921820] env[63355]: DEBUG nova.compute.utils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Binding failed for port 6038aaf0-9377-44f3-adc1-bb4c25210d18, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.925202] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.899s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.925202] env[63355]: INFO nova.compute.claims [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.928651] env[63355]: DEBUG nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Build of instance 35bd361f-fda1-4121-9f4c-61d3219f9202 was re-scheduled: Binding failed for port 6038aaf0-9377-44f3-adc1-bb4c25210d18, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.929159] env[63355]: DEBUG nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.929267] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Acquiring lock "refresh_cache-35bd361f-fda1-4121-9f4c-61d3219f9202" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.933676] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Acquired lock "refresh_cache-35bd361f-fda1-4121-9f4c-61d3219f9202" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.933676] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.949435] env[63355]: DEBUG oslo_vmware.rw_handles [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63355) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 598.178944] env[63355]: DEBUG nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 598.227018] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 598.227018] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 598.227018] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 598.228030] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 598.228030] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 598.228030] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 598.228497] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 598.229311] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 598.229311] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 598.229555] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 598.229866] env[63355]: DEBUG nova.virt.hardware [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.231493] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96cc65b-b4e8-40b3-9438-93675be0fddb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.246341] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3238149-9fec-4d6e-bce5-d929d367f7b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.266562] env[63355]: DEBUG nova.network.neutron [req-45135d48-7609-42aa-b81d-173fbb2d9abe req-9b75763e-f912-4ac4-bcfb-ad3af9f1a104 service nova] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.317344] env[63355]: DEBUG nova.network.neutron [-] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.612361] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.769964] env[63355]: DEBUG nova.network.neutron [req-45135d48-7609-42aa-b81d-173fbb2d9abe req-9b75763e-f912-4ac4-bcfb-ad3af9f1a104 service nova] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.823327] env[63355]: DEBUG oslo_vmware.rw_handles [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Completed reading data from the image iterator. {{(pid=63355) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 598.823487] env[63355]: DEBUG oslo_vmware.rw_handles [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 598.824051] env[63355]: INFO nova.compute.manager [-] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Took 1.05 seconds to deallocate network for instance. [ 598.829433] env[63355]: DEBUG nova.compute.claims [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 598.829652] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.914064] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.960952] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Downloaded image file data 84ca0bb3-9916-4cea-9399-75af77b0558e to vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk on the data store datastore2 {{(pid=63355) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 598.963056] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Caching image {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 598.963328] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Copying Virtual Disk [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk to [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 598.963597] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a3d1dae-1a8e-48ae-b924-b544ad3d35d0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.975033] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 598.975033] env[63355]: value = "task-1349464" [ 598.975033] env[63355]: _type = "Task" [ 598.975033] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.987254] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349464, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.133510] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b56eee6-3d55-4e52-93b3-23edeacc05fe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.141571] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4768559a-ca70-4e44-82c7-c0b1d480f7ed {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.185531] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32482de-fd2c-4d3f-a814-62dd35f036d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.194510] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a86e70-692e-4a5c-800a-7aab6d1a73f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.209589] env[63355]: DEBUG nova.compute.provider_tree [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.276285] env[63355]: DEBUG oslo_concurrency.lockutils [req-45135d48-7609-42aa-b81d-173fbb2d9abe req-9b75763e-f912-4ac4-bcfb-ad3af9f1a104 service nova] Releasing lock "refresh_cache-ce2f828a-82d0-4270-ac21-d967b2e0594b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.417086] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Releasing lock "refresh_cache-35bd361f-fda1-4121-9f4c-61d3219f9202" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.417086] env[63355]: DEBUG nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 599.417386] env[63355]: DEBUG nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.417665] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.446738] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.487706] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349464, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.567187] env[63355]: ERROR nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 29895cad-8791-4e5d-8afc-ea7cd1027834, please check neutron logs for more information. [ 599.567187] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 599.567187] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.567187] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 599.567187] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.567187] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 599.567187] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.567187] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 599.567187] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.567187] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 599.567187] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.567187] env[63355]: ERROR nova.compute.manager raise self.value [ 599.567187] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.567187] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 599.567187] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.567187] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 599.567675] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.567675] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 599.567675] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 29895cad-8791-4e5d-8afc-ea7cd1027834, please check neutron logs for more information. [ 599.567675] env[63355]: ERROR nova.compute.manager [ 599.567675] env[63355]: Traceback (most recent call last): [ 599.567675] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 599.567675] env[63355]: listener.cb(fileno) [ 599.567675] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.567675] env[63355]: result = function(*args, **kwargs) [ 599.567675] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.567675] env[63355]: return func(*args, **kwargs) [ 599.567675] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.567675] env[63355]: raise e [ 599.567675] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.567675] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 599.567675] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.567675] env[63355]: created_port_ids = self._update_ports_for_instance( [ 599.567675] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.567675] env[63355]: with excutils.save_and_reraise_exception(): [ 599.567675] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.567675] env[63355]: self.force_reraise() [ 599.567675] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.567675] env[63355]: raise self.value [ 599.567675] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.567675] env[63355]: updated_port = self._update_port( [ 599.567675] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.567675] env[63355]: _ensure_no_port_binding_failure(port) [ 599.567675] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.567675] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.568512] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 29895cad-8791-4e5d-8afc-ea7cd1027834, please check neutron logs for more information. [ 599.568512] env[63355]: Removing descriptor: 18 [ 599.568512] env[63355]: ERROR nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 29895cad-8791-4e5d-8afc-ea7cd1027834, please check neutron logs for more information. [ 599.568512] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Traceback (most recent call last): [ 599.568512] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 599.568512] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] yield resources [ 599.568512] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.568512] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] self.driver.spawn(context, instance, image_meta, [ 599.568512] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.568512] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.568512] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.568512] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] vm_ref = self.build_virtual_machine(instance, [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] for vif in network_info: [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] return self._sync_wrapper(fn, *args, **kwargs) [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] self.wait() [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] self[:] = self._gt.wait() [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] return self._exit_event.wait() [ 599.568835] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] result = hub.switch() [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] return self.greenlet.switch() [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] result = function(*args, **kwargs) [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] return func(*args, **kwargs) [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] raise e [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] nwinfo = self.network_api.allocate_for_instance( [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.569274] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] created_port_ids = self._update_ports_for_instance( [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] with excutils.save_and_reraise_exception(): [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] self.force_reraise() [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] raise self.value [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] updated_port = self._update_port( [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] _ensure_no_port_binding_failure(port) [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.569662] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] raise exception.PortBindingFailed(port_id=port['id']) [ 599.569957] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] nova.exception.PortBindingFailed: Binding failed for port 29895cad-8791-4e5d-8afc-ea7cd1027834, please check neutron logs for more information. [ 599.569957] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] [ 599.569957] env[63355]: INFO nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Terminating instance [ 599.572809] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Acquiring lock "refresh_cache-8d2ea6a5-5c57-468f-be08-62f418b80e1e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.573056] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Acquired lock "refresh_cache-8d2ea6a5-5c57-468f-be08-62f418b80e1e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.573360] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 599.712868] env[63355]: DEBUG nova.scheduler.client.report [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.950897] env[63355]: DEBUG nova.network.neutron [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.989598] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349464, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.713146} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.989859] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Copied Virtual Disk [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk to [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 599.992439] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Deleting the datastore file [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 599.992825] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4adc98f0-d6b2-4b01-adb3-7e9f86198cc5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.002295] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 600.002295] env[63355]: value = "task-1349465" [ 600.002295] env[63355]: _type = "Task" [ 600.002295] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.010848] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349465, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.123110] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.220273] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.295s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.220273] env[63355]: DEBUG nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.223536] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.598s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.364606] env[63355]: DEBUG nova.compute.manager [req-f944afc1-139b-4b0d-8dea-ce5df33b7d13 req-cb460392-b371-45b0-92c7-0a5378c00faa service nova] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Received event network-vif-deleted-f0619064-8c90-4914-9df4-a1e6286d3d89 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.406141] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.454106] env[63355]: INFO nova.compute.manager [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] Took 1.04 seconds to deallocate network for instance. [ 600.520522] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349465, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023486} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.520806] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 600.521039] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Moving file from [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122/84ca0bb3-9916-4cea-9399-75af77b0558e to [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e. {{(pid=63355) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 600.521678] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-ca5f9157-6150-4bf9-b191-796057b64960 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.530663] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 600.530663] env[63355]: value = "task-1349466" [ 600.530663] env[63355]: _type = "Task" [ 600.530663] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.540093] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349466, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.736486] env[63355]: DEBUG nova.compute.utils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.738471] env[63355]: DEBUG nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.738471] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 600.887165] env[63355]: DEBUG nova.policy [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17454180659b4662b564de0b390738f7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0fadcfa825ef4cb7b6ca83b6612440ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.912818] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Releasing lock "refresh_cache-8d2ea6a5-5c57-468f-be08-62f418b80e1e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.914226] env[63355]: DEBUG nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 600.914488] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 600.914839] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c77f2ee-57b8-413b-a5e3-3fe6c9a302db {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.935293] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22834194-4f5a-4580-bf36-9290e2a7afac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.972998] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8d2ea6a5-5c57-468f-be08-62f418b80e1e could not be found. [ 600.973276] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 600.973384] env[63355]: INFO nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Took 0.06 seconds to destroy the instance on the hypervisor. [ 600.973627] env[63355]: DEBUG oslo.service.loopingcall [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.976330] env[63355]: DEBUG nova.compute.manager [-] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.976424] env[63355]: DEBUG nova.network.neutron [-] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.001699] env[63355]: DEBUG nova.network.neutron [-] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.015902] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0034e262-3065-436b-911a-1310645a59fc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.028242] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5b232f-42c5-4ae7-8343-4ada8077ff17 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.046971] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349466, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024036} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.081148] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] File moved {{(pid=63355) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 601.081148] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Cleaning up location [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122 {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 601.081148] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Deleting the datastore file [datastore2] vmware_temp/f01a948e-6323-4d17-aff0-78b61d9cb122 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 601.081148] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c826fc16-f239-4efd-81bb-03d070e62095 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.086141] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb21541-82e8-422a-b4c0-279fe9168bd5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.092812] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f2c95c-6298-41bc-91f5-15d41f02ff12 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.097193] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 601.097193] env[63355]: value = "task-1349467" [ 601.097193] env[63355]: _type = "Task" [ 601.097193] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.108735] env[63355]: DEBUG nova.compute.provider_tree [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.122319] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349467, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.122651] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Acquiring lock "c80d746b-ad37-43d7-874e-1c1c848546f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.122909] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Lock "c80d746b-ad37-43d7-874e-1c1c848546f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.221073] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Acquiring lock "591a7907-e4a7-41d6-9cd0-5cb359f08810" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.221300] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Lock "591a7907-e4a7-41d6-9cd0-5cb359f08810" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.242528] env[63355]: DEBUG nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.482249] env[63355]: INFO nova.scheduler.client.report [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Deleted allocations for instance 35bd361f-fda1-4121-9f4c-61d3219f9202 [ 601.505561] env[63355]: DEBUG nova.network.neutron [-] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.615154] env[63355]: DEBUG nova.scheduler.client.report [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.618404] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349467, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027355} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.620636] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 601.620820] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b22037b-2e22-440d-9145-1e6d2f74b2db {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.630662] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 601.630662] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c9f9ba-7010-eea5-9a6e-0419058fcf53" [ 601.630662] env[63355]: _type = "Task" [ 601.630662] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.647025] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c9f9ba-7010-eea5-9a6e-0419058fcf53, 'name': SearchDatastore_Task, 'duration_secs': 0.009361} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.647025] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.647025] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] abe6726e-7d6a-46e0-9273-de3e681311eb/abe6726e-7d6a-46e0-9273-de3e681311eb.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 601.647025] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-694c0b9c-127d-4a4d-9a12-b5828d43ace0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.656223] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 601.656223] env[63355]: value = "task-1349468" [ 601.656223] env[63355]: _type = "Task" [ 601.656223] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.663946] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349468, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.958658] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "def90353-ae90-4e1e-9fe1-8e5459a7309d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.959410] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "def90353-ae90-4e1e-9fe1-8e5459a7309d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.000805] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0e330539-81a4-4bb8-a030-3139c1b7f8fd tempest-ImagesOneServerTestJSON-666055249 tempest-ImagesOneServerTestJSON-666055249-project-member] Lock "35bd361f-fda1-4121-9f4c-61d3219f9202" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.712s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.002295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "35bd361f-fda1-4121-9f4c-61d3219f9202" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.338s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.002724] env[63355]: INFO nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 35bd361f-fda1-4121-9f4c-61d3219f9202] During sync_power_state the instance has a pending task (spawning). Skip. [ 602.002724] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "35bd361f-fda1-4121-9f4c-61d3219f9202" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.009989] env[63355]: INFO nova.compute.manager [-] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Took 1.03 seconds to deallocate network for instance. [ 602.020040] env[63355]: DEBUG nova.compute.claims [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 602.020040] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.121558] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.899s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.122264] env[63355]: ERROR nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d96025ae-f12d-493d-90c3-8d3fefd523e4, please check neutron logs for more information. [ 602.122264] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Traceback (most recent call last): [ 602.122264] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.122264] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] self.driver.spawn(context, instance, image_meta, [ 602.122264] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.122264] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.122264] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.122264] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] vm_ref = self.build_virtual_machine(instance, [ 602.122264] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.122264] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.122264] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] for vif in network_info: [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] return self._sync_wrapper(fn, *args, **kwargs) [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] self.wait() [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] self[:] = self._gt.wait() [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] return self._exit_event.wait() [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] result = hub.switch() [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.122594] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] return self.greenlet.switch() [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] result = function(*args, **kwargs) [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] return func(*args, **kwargs) [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] raise e [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] nwinfo = self.network_api.allocate_for_instance( [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] created_port_ids = self._update_ports_for_instance( [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] with excutils.save_and_reraise_exception(): [ 602.123129] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] self.force_reraise() [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] raise self.value [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] updated_port = self._update_port( [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] _ensure_no_port_binding_failure(port) [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] raise exception.PortBindingFailed(port_id=port['id']) [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] nova.exception.PortBindingFailed: Binding failed for port d96025ae-f12d-493d-90c3-8d3fefd523e4, please check neutron logs for more information. [ 602.123483] env[63355]: ERROR nova.compute.manager [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] [ 602.123857] env[63355]: DEBUG nova.compute.utils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Binding failed for port d96025ae-f12d-493d-90c3-8d3fefd523e4, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.124728] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.683s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.130168] env[63355]: DEBUG nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Build of instance b3d31162-5851-466f-b44c-d15e69c76bdd was re-scheduled: Binding failed for port d96025ae-f12d-493d-90c3-8d3fefd523e4, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 602.130168] env[63355]: DEBUG nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 602.130168] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquiring lock "refresh_cache-b3d31162-5851-466f-b44c-d15e69c76bdd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.130168] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquired lock "refresh_cache-b3d31162-5851-466f-b44c-d15e69c76bdd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.130416] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.166902] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349468, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.248623] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Acquiring lock "4392313f-2fed-4308-b79a-324b3431d5af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.248860] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Lock "4392313f-2fed-4308-b79a-324b3431d5af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.255908] env[63355]: DEBUG nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.300105] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.300299] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.300466] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.300657] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.300833] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.300982] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.301212] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.301367] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.301565] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.301734] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.301909] env[63355]: DEBUG nova.virt.hardware [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.302856] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3858a98-64fb-477c-80dd-96e987e2658c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.320587] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25de5f73-6b24-4ee2-88a7-ad7d69c1226b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.362195] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Successfully created port: 0423d0dc-4858-4959-91a9-67573bda0806 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.504402] env[63355]: DEBUG nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 602.672031] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349468, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51492} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.672031] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] abe6726e-7d6a-46e0-9273-de3e681311eb/abe6726e-7d6a-46e0-9273-de3e681311eb.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 602.672225] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 602.672625] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d49b7da8-c123-4835-a639-89cc7f155c51 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.680402] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 602.680402] env[63355]: value = "task-1349469" [ 602.680402] env[63355]: _type = "Task" [ 602.680402] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.698749] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349469, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.731672] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.959242] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e6a775-d2e6-4130-b6d6-188eb9e0ee4f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.967587] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b951aba8-7901-4f4f-8bf1-58c8114356c8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.975371] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Acquiring lock "bc1d7402-3651-488b-a431-2cb2ae987d32" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.975627] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Lock "bc1d7402-3651-488b-a431-2cb2ae987d32" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.002032] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523aaaf5-c321-4dc5-9ec7-213507f08048 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.009689] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4cac07-54d5-4c89-83fe-89c5eb53b1b7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.017446] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.029785] env[63355]: DEBUG nova.compute.provider_tree [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.057296] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.065781] env[63355]: DEBUG nova.compute.manager [req-ec927735-0425-4336-a02b-9f7d12af50b1 req-21e4f685-a01c-4cb4-be64-6ba7c9970e29 service nova] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Received event network-changed-29895cad-8791-4e5d-8afc-ea7cd1027834 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.066037] env[63355]: DEBUG nova.compute.manager [req-ec927735-0425-4336-a02b-9f7d12af50b1 req-21e4f685-a01c-4cb4-be64-6ba7c9970e29 service nova] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Refreshing instance network info cache due to event network-changed-29895cad-8791-4e5d-8afc-ea7cd1027834. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 603.066195] env[63355]: DEBUG oslo_concurrency.lockutils [req-ec927735-0425-4336-a02b-9f7d12af50b1 req-21e4f685-a01c-4cb4-be64-6ba7c9970e29 service nova] Acquiring lock "refresh_cache-8d2ea6a5-5c57-468f-be08-62f418b80e1e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.066341] env[63355]: DEBUG oslo_concurrency.lockutils [req-ec927735-0425-4336-a02b-9f7d12af50b1 req-21e4f685-a01c-4cb4-be64-6ba7c9970e29 service nova] Acquired lock "refresh_cache-8d2ea6a5-5c57-468f-be08-62f418b80e1e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.066481] env[63355]: DEBUG nova.network.neutron [req-ec927735-0425-4336-a02b-9f7d12af50b1 req-21e4f685-a01c-4cb4-be64-6ba7c9970e29 service nova] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Refreshing network info cache for port 29895cad-8791-4e5d-8afc-ea7cd1027834 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 603.194560] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349469, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067746} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.194913] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 603.196649] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ac8cf0-6748-4284-9b8a-11d331471cb4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.226617] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] abe6726e-7d6a-46e0-9273-de3e681311eb/abe6726e-7d6a-46e0-9273-de3e681311eb.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 603.227144] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56dbd74d-e192-4e04-becc-a1f9589ba547 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.250007] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 603.250007] env[63355]: value = "task-1349470" [ 603.250007] env[63355]: _type = "Task" [ 603.250007] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.257960] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349470, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.521105] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Releasing lock "refresh_cache-b3d31162-5851-466f-b44c-d15e69c76bdd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.521412] env[63355]: DEBUG nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 603.521519] env[63355]: DEBUG nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.521677] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.532981] env[63355]: DEBUG nova.scheduler.client.report [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.574046] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.760717] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349470, 'name': ReconfigVM_Task, 'duration_secs': 0.281767} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.760937] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Reconfigured VM instance instance-00000007 to attach disk [datastore2] abe6726e-7d6a-46e0-9273-de3e681311eb/abe6726e-7d6a-46e0-9273-de3e681311eb.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 603.762766] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0de6969c-ee8a-4347-b26a-9ec46f02661e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.771032] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 603.771032] env[63355]: value = "task-1349471" [ 603.771032] env[63355]: _type = "Task" [ 603.771032] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.777584] env[63355]: DEBUG nova.network.neutron [req-ec927735-0425-4336-a02b-9f7d12af50b1 req-21e4f685-a01c-4cb4-be64-6ba7c9970e29 service nova] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.791467] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349471, 'name': Rename_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.044011] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.916s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.045966] env[63355]: ERROR nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 16b85a19-bea8-4470-82d6-d4c17ade8cf1, please check neutron logs for more information. [ 604.045966] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Traceback (most recent call last): [ 604.045966] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.045966] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] self.driver.spawn(context, instance, image_meta, [ 604.045966] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 604.045966] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.045966] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.045966] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] vm_ref = self.build_virtual_machine(instance, [ 604.045966] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.045966] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.045966] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] for vif in network_info: [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] return self._sync_wrapper(fn, *args, **kwargs) [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] self.wait() [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] self[:] = self._gt.wait() [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] return self._exit_event.wait() [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] result = hub.switch() [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.046337] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] return self.greenlet.switch() [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] result = function(*args, **kwargs) [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] return func(*args, **kwargs) [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] raise e [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] nwinfo = self.network_api.allocate_for_instance( [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] created_port_ids = self._update_ports_for_instance( [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] with excutils.save_and_reraise_exception(): [ 604.046667] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] self.force_reraise() [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] raise self.value [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] updated_port = self._update_port( [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] _ensure_no_port_binding_failure(port) [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] raise exception.PortBindingFailed(port_id=port['id']) [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] nova.exception.PortBindingFailed: Binding failed for port 16b85a19-bea8-4470-82d6-d4c17ade8cf1, please check neutron logs for more information. [ 604.047330] env[63355]: ERROR nova.compute.manager [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] [ 604.047618] env[63355]: DEBUG nova.compute.utils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Binding failed for port 16b85a19-bea8-4470-82d6-d4c17ade8cf1, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.050959] env[63355]: DEBUG nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Build of instance d08c461b-c2ba-4484-a0aa-746d726868a0 was re-scheduled: Binding failed for port 16b85a19-bea8-4470-82d6-d4c17ade8cf1, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.050959] env[63355]: DEBUG nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.050959] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Acquiring lock "refresh_cache-d08c461b-c2ba-4484-a0aa-746d726868a0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.050959] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Acquired lock "refresh_cache-d08c461b-c2ba-4484-a0aa-746d726868a0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.051191] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.057680] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.358s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.070505] env[63355]: DEBUG nova.network.neutron [req-ec927735-0425-4336-a02b-9f7d12af50b1 req-21e4f685-a01c-4cb4-be64-6ba7c9970e29 service nova] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.076868] env[63355]: DEBUG nova.network.neutron [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.283595] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349471, 'name': Rename_Task, 'duration_secs': 0.127991} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.283787] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 604.284083] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24919876-32bd-447d-963f-5c24f7268e2d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.294449] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 604.294449] env[63355]: value = "task-1349472" [ 604.294449] env[63355]: _type = "Task" [ 604.294449] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.307890] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349472, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.560283] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "f9c8ed86-506b-4654-8e7e-cb218605764b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.560575] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "f9c8ed86-506b-4654-8e7e-cb218605764b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.574087] env[63355]: DEBUG oslo_concurrency.lockutils [req-ec927735-0425-4336-a02b-9f7d12af50b1 req-21e4f685-a01c-4cb4-be64-6ba7c9970e29 service nova] Releasing lock "refresh_cache-8d2ea6a5-5c57-468f-be08-62f418b80e1e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.574087] env[63355]: DEBUG nova.compute.manager [req-ec927735-0425-4336-a02b-9f7d12af50b1 req-21e4f685-a01c-4cb4-be64-6ba7c9970e29 service nova] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Received event network-vif-deleted-29895cad-8791-4e5d-8afc-ea7cd1027834 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.580318] env[63355]: INFO nova.compute.manager [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] Took 1.06 seconds to deallocate network for instance. [ 604.669168] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.815255] env[63355]: DEBUG oslo_vmware.api [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349472, 'name': PowerOnVM_Task, 'duration_secs': 0.417497} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.822011] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 604.822608] env[63355]: INFO nova.compute.manager [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Took 8.94 seconds to spawn the instance on the hypervisor. [ 604.822608] env[63355]: DEBUG nova.compute.manager [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 604.824348] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870a9bdb-ae10-4250-97fa-82501e84777c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.953263] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4618d28c-4f1f-418e-848d-5a0391cb0260 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.966118] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ee087f-5608-4f6a-a1fa-491f1e83b696 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.999272] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2455df6b-c1ed-433d-81cc-5f6e75277218 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.007484] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9f599a-a8fd-4981-b27c-25972c312fc0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.025732] env[63355]: DEBUG nova.compute.provider_tree [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.275265] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.356980] env[63355]: INFO nova.compute.manager [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Took 17.73 seconds to build instance. [ 605.530042] env[63355]: DEBUG nova.scheduler.client.report [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.629183] env[63355]: INFO nova.scheduler.client.report [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Deleted allocations for instance b3d31162-5851-466f-b44c-d15e69c76bdd [ 605.778625] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Releasing lock "refresh_cache-d08c461b-c2ba-4484-a0aa-746d726868a0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.778848] env[63355]: DEBUG nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.779049] env[63355]: DEBUG nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.779215] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.858438] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6beeba79-3d05-4ffa-824d-f19a24299568 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lock "abe6726e-7d6a-46e0-9273-de3e681311eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.248s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.863497] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "abe6726e-7d6a-46e0-9273-de3e681311eb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 10.195s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.863497] env[63355]: INFO nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] During sync_power_state the instance has a pending task (block_device_mapping). Skip. [ 605.863497] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "abe6726e-7d6a-46e0-9273-de3e681311eb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.865046] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.043671] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.044609] env[63355]: ERROR nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f270dd28-d082-4021-8f15-d71f9eec8b01, please check neutron logs for more information. [ 606.044609] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Traceback (most recent call last): [ 606.044609] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.044609] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] self.driver.spawn(context, instance, image_meta, [ 606.044609] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 606.044609] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.044609] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.044609] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] vm_ref = self.build_virtual_machine(instance, [ 606.044609] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.044609] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.044609] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] for vif in network_info: [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] return self._sync_wrapper(fn, *args, **kwargs) [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] self.wait() [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] self[:] = self._gt.wait() [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] return self._exit_event.wait() [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] result = hub.switch() [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.044944] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] return self.greenlet.switch() [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] result = function(*args, **kwargs) [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] return func(*args, **kwargs) [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] raise e [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] nwinfo = self.network_api.allocate_for_instance( [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] created_port_ids = self._update_ports_for_instance( [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] with excutils.save_and_reraise_exception(): [ 606.045281] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] self.force_reraise() [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] raise self.value [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] updated_port = self._update_port( [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] _ensure_no_port_binding_failure(port) [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] raise exception.PortBindingFailed(port_id=port['id']) [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] nova.exception.PortBindingFailed: Binding failed for port f270dd28-d082-4021-8f15-d71f9eec8b01, please check neutron logs for more information. [ 606.045610] env[63355]: ERROR nova.compute.manager [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] [ 606.045897] env[63355]: DEBUG nova.compute.utils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Binding failed for port f270dd28-d082-4021-8f15-d71f9eec8b01, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 606.045897] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.680s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.050044] env[63355]: INFO nova.compute.claims [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.052240] env[63355]: DEBUG nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Build of instance 059dc72c-dc91-4744-8efa-886aef99d5c5 was re-scheduled: Binding failed for port f270dd28-d082-4021-8f15-d71f9eec8b01, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 606.052913] env[63355]: DEBUG nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 606.052913] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquiring lock "refresh_cache-059dc72c-dc91-4744-8efa-886aef99d5c5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.053058] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquired lock "refresh_cache-059dc72c-dc91-4744-8efa-886aef99d5c5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.054767] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.150802] env[63355]: DEBUG oslo_concurrency.lockutils [None req-abbe0865-19ad-441d-ae0a-a5f5accce0bc tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "b3d31162-5851-466f-b44c-d15e69c76bdd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.765s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.151450] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "b3d31162-5851-466f-b44c-d15e69c76bdd" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 10.487s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.151792] env[63355]: INFO nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: b3d31162-5851-466f-b44c-d15e69c76bdd] During sync_power_state the instance has a pending task (spawning). Skip. [ 606.152042] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "b3d31162-5851-466f-b44c-d15e69c76bdd" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.351119] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Acquiring lock "1de0e894-1edb-4296-9e9b-2c1eb0c7c275" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.351119] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Lock "1de0e894-1edb-4296-9e9b-2c1eb0c7c275" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.362500] env[63355]: DEBUG nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 606.368786] env[63355]: DEBUG nova.network.neutron [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.605649] env[63355]: ERROR nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0423d0dc-4858-4959-91a9-67573bda0806, please check neutron logs for more information. [ 606.605649] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.605649] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.605649] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.605649] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.605649] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.605649] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.605649] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.605649] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.605649] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 606.605649] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.605649] env[63355]: ERROR nova.compute.manager raise self.value [ 606.605649] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.605649] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.605649] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.605649] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.606279] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.606279] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.606279] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0423d0dc-4858-4959-91a9-67573bda0806, please check neutron logs for more information. [ 606.606279] env[63355]: ERROR nova.compute.manager [ 606.606279] env[63355]: Traceback (most recent call last): [ 606.606279] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.606279] env[63355]: listener.cb(fileno) [ 606.606279] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.606279] env[63355]: result = function(*args, **kwargs) [ 606.606279] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.606279] env[63355]: return func(*args, **kwargs) [ 606.606279] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.606279] env[63355]: raise e [ 606.606279] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.606279] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 606.606279] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.606279] env[63355]: created_port_ids = self._update_ports_for_instance( [ 606.606279] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.606279] env[63355]: with excutils.save_and_reraise_exception(): [ 606.606279] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.606279] env[63355]: self.force_reraise() [ 606.606279] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.606279] env[63355]: raise self.value [ 606.606279] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.606279] env[63355]: updated_port = self._update_port( [ 606.606279] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.606279] env[63355]: _ensure_no_port_binding_failure(port) [ 606.606279] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.606279] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.607059] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 0423d0dc-4858-4959-91a9-67573bda0806, please check neutron logs for more information. [ 606.607059] env[63355]: Removing descriptor: 18 [ 606.607059] env[63355]: ERROR nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0423d0dc-4858-4959-91a9-67573bda0806, please check neutron logs for more information. [ 606.607059] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Traceback (most recent call last): [ 606.607059] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.607059] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] yield resources [ 606.607059] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.607059] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] self.driver.spawn(context, instance, image_meta, [ 606.607059] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 606.607059] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.607059] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.607059] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] vm_ref = self.build_virtual_machine(instance, [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] for vif in network_info: [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] return self._sync_wrapper(fn, *args, **kwargs) [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] self.wait() [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] self[:] = self._gt.wait() [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] return self._exit_event.wait() [ 606.607516] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] result = hub.switch() [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] return self.greenlet.switch() [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] result = function(*args, **kwargs) [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] return func(*args, **kwargs) [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] raise e [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] nwinfo = self.network_api.allocate_for_instance( [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.608768] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] created_port_ids = self._update_ports_for_instance( [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] with excutils.save_and_reraise_exception(): [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] self.force_reraise() [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] raise self.value [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] updated_port = self._update_port( [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] _ensure_no_port_binding_failure(port) [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.609139] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] raise exception.PortBindingFailed(port_id=port['id']) [ 606.609434] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] nova.exception.PortBindingFailed: Binding failed for port 0423d0dc-4858-4959-91a9-67573bda0806, please check neutron logs for more information. [ 606.609434] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] [ 606.609434] env[63355]: INFO nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Terminating instance [ 606.609434] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Acquiring lock "refresh_cache-0fb6ba21-d45e-4809-8e68-d78a5892baeb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.609434] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Acquired lock "refresh_cache-0fb6ba21-d45e-4809-8e68-d78a5892baeb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.609577] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.627964] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.658860] env[63355]: DEBUG nova.compute.manager [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 606.823200] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.876723] env[63355]: INFO nova.compute.manager [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] Took 1.10 seconds to deallocate network for instance. [ 606.895210] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.184191] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.192867] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.329864] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Releasing lock "refresh_cache-059dc72c-dc91-4744-8efa-886aef99d5c5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.330317] env[63355]: DEBUG nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 607.330600] env[63355]: DEBUG nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.331067] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.373507] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3fd7af-b473-4d89-8f8b-2bf8302bb86e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.394234] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bb272e-847e-4fe7-9173-2b4b16f8e8ee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.400650] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.435661] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c9b237-bc34-42ed-9759-01dea2aae729 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.448829] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c518f9-2e1f-4c93-a26e-efa85cf88c24 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.463468] env[63355]: DEBUG nova.compute.provider_tree [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.635507] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.906451] env[63355]: DEBUG nova.network.neutron [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.936086] env[63355]: DEBUG nova.compute.manager [req-7f9c44f9-d451-48ea-b71d-3e427f4c257c req-a204c66f-93b6-4e05-821a-18d3d8226f05 service nova] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Received event network-changed-0423d0dc-4858-4959-91a9-67573bda0806 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 607.936086] env[63355]: DEBUG nova.compute.manager [req-7f9c44f9-d451-48ea-b71d-3e427f4c257c req-a204c66f-93b6-4e05-821a-18d3d8226f05 service nova] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Refreshing instance network info cache due to event network-changed-0423d0dc-4858-4959-91a9-67573bda0806. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 607.936086] env[63355]: DEBUG oslo_concurrency.lockutils [req-7f9c44f9-d451-48ea-b71d-3e427f4c257c req-a204c66f-93b6-4e05-821a-18d3d8226f05 service nova] Acquiring lock "refresh_cache-0fb6ba21-d45e-4809-8e68-d78a5892baeb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.940216] env[63355]: INFO nova.scheduler.client.report [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Deleted allocations for instance d08c461b-c2ba-4484-a0aa-746d726868a0 [ 607.968518] env[63355]: DEBUG nova.scheduler.client.report [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.121146] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Acquiring lock "10112b62-f41e-4413-9bc8-c1306c5c1bc1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.121652] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Lock "10112b62-f41e-4413-9bc8-c1306c5c1bc1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.141938] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Releasing lock "refresh_cache-0fb6ba21-d45e-4809-8e68-d78a5892baeb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.142386] env[63355]: DEBUG nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.142572] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 608.142950] env[63355]: DEBUG oslo_concurrency.lockutils [req-7f9c44f9-d451-48ea-b71d-3e427f4c257c req-a204c66f-93b6-4e05-821a-18d3d8226f05 service nova] Acquired lock "refresh_cache-0fb6ba21-d45e-4809-8e68-d78a5892baeb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.143590] env[63355]: DEBUG nova.network.neutron [req-7f9c44f9-d451-48ea-b71d-3e427f4c257c req-a204c66f-93b6-4e05-821a-18d3d8226f05 service nova] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Refreshing network info cache for port 0423d0dc-4858-4959-91a9-67573bda0806 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 608.144518] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f740c44c-510c-41ed-bafa-2a854625d968 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.155900] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe5d836-2e50-4d17-8d5f-4c3e42ec08ee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.184884] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0fb6ba21-d45e-4809-8e68-d78a5892baeb could not be found. [ 608.185146] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 608.185336] env[63355]: INFO nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 608.185575] env[63355]: DEBUG oslo.service.loopingcall [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.185791] env[63355]: DEBUG nova.compute.manager [-] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.185880] env[63355]: DEBUG nova.network.neutron [-] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 608.213498] env[63355]: DEBUG nova.network.neutron [-] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.409913] env[63355]: INFO nova.compute.manager [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] Took 1.08 seconds to deallocate network for instance. [ 608.451872] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa84c6fd-6192-4e86-9b08-47fd88d6a512 tempest-AttachInterfacesUnderV243Test-1786759218 tempest-AttachInterfacesUnderV243Test-1786759218-project-member] Lock "d08c461b-c2ba-4484-a0aa-746d726868a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.814s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.454157] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "d08c461b-c2ba-4484-a0aa-746d726868a0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 12.789s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.454157] env[63355]: INFO nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: d08c461b-c2ba-4484-a0aa-746d726868a0] During sync_power_state the instance has a pending task (spawning). Skip. [ 608.454157] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "d08c461b-c2ba-4484-a0aa-746d726868a0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.474713] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.475698] env[63355]: DEBUG nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.478418] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.660s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.480825] env[63355]: INFO nova.compute.claims [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.670034] env[63355]: DEBUG nova.network.neutron [req-7f9c44f9-d451-48ea-b71d-3e427f4c257c req-a204c66f-93b6-4e05-821a-18d3d8226f05 service nova] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.716466] env[63355]: DEBUG nova.network.neutron [-] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.900704] env[63355]: DEBUG nova.network.neutron [req-7f9c44f9-d451-48ea-b71d-3e427f4c257c req-a204c66f-93b6-4e05-821a-18d3d8226f05 service nova] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.959837] env[63355]: DEBUG nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.988300] env[63355]: DEBUG nova.compute.utils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.992539] env[63355]: DEBUG nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 608.992539] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 609.106416] env[63355]: DEBUG nova.policy [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b36cc59df054695ae7fbe284ba3d5b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08b7ec69fbd14616828658200e33bd32', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 609.220658] env[63355]: INFO nova.compute.manager [-] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Took 1.03 seconds to deallocate network for instance. [ 609.226411] env[63355]: DEBUG nova.compute.claims [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 609.226768] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.403035] env[63355]: DEBUG oslo_concurrency.lockutils [req-7f9c44f9-d451-48ea-b71d-3e427f4c257c req-a204c66f-93b6-4e05-821a-18d3d8226f05 service nova] Releasing lock "refresh_cache-0fb6ba21-d45e-4809-8e68-d78a5892baeb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.440333] env[63355]: INFO nova.scheduler.client.report [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Deleted allocations for instance 059dc72c-dc91-4744-8efa-886aef99d5c5 [ 609.491313] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.493315] env[63355]: DEBUG nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.721404] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Successfully created port: 34286625-39c6-4641-9f1e-99ec49be4177 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.784496] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025fecf2-f38c-4615-8307-039bba271a7e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.793578] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa01bcc-5013-4c72-abb4-8d73f09722a0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.832557] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba141f57-820a-4213-984a-78d1e744f774 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.843941] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5723790f-a8c3-49b8-b64c-e5672a281569 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.861366] env[63355]: DEBUG nova.compute.provider_tree [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.904945] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Acquiring lock "b275344f-695e-463b-8bfe-2bee97296349" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.905255] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Lock "b275344f-695e-463b-8bfe-2bee97296349" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.952817] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9a8228f-58d5-4b98-9f5c-4347a739b0e3 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "059dc72c-dc91-4744-8efa-886aef99d5c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.464s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.956890] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "059dc72c-dc91-4744-8efa-886aef99d5c5" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 14.289s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.956890] env[63355]: INFO nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 059dc72c-dc91-4744-8efa-886aef99d5c5] During sync_power_state the instance has a pending task (spawning). Skip. [ 609.956890] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "059dc72c-dc91-4744-8efa-886aef99d5c5" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.344842] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Acquiring lock "3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.345130] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Lock "3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.365216] env[63355]: DEBUG nova.scheduler.client.report [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.456362] env[63355]: DEBUG nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 610.509698] env[63355]: DEBUG nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.550755] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.550755] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.550755] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.550755] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.551198] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.552727] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.553016] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.553186] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.553351] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.553519] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.553690] env[63355]: DEBUG nova.virt.hardware [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.556152] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe50091-903a-4a77-abcb-8ab16914444e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.570395] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01175e71-432c-48d6-acfe-776cc20316d6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.672212] env[63355]: DEBUG nova.compute.manager [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 610.672733] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92403711-4aed-4d51-a37c-000295a075a7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.873208] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.873910] env[63355]: DEBUG nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 610.879344] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.710s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.879658] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.882050] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 610.882050] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.051s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.885183] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc7dba1-fc69-43aa-9350-29c211caabed {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.897244] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20afc5e6-c9bb-4513-a6d5-7c62e5d9f705 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.920663] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e86d46-ec94-40a0-881c-64aca46aa3e0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.931444] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c65f76-8bc5-4111-a2a7-739c59dfe497 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.987452] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181504MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 610.987617] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.014978] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.184397] env[63355]: INFO nova.compute.manager [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] instance snapshotting [ 611.185099] env[63355]: DEBUG nova.objects.instance [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lazy-loading 'flavor' on Instance uuid abe6726e-7d6a-46e0-9273-de3e681311eb {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 611.388058] env[63355]: DEBUG nova.compute.utils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.389479] env[63355]: DEBUG nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 611.390740] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 611.541844] env[63355]: DEBUG nova.policy [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c74e36da1fe487bb8990981a8124757', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89579fa6fc9c457a87177d3b64b7215b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 611.582599] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "0de6892a-5463-4d60-807c-0e3fca2accf3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.582930] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "0de6892a-5463-4d60-807c-0e3fca2accf3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.691929] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224b47a5-67dc-4e95-8526-5fee3c02e0c5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.700855] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4f9dcf-2c12-4775-99f9-a831414c8981 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.728189] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29802505-d957-48db-9c5c-56ccd97d4ae2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.736675] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e31a55-9d1e-45af-a68a-b21d05f6b653 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.771848] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a79ff56-fb55-4b65-b7d8-206485267172 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.780734] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e89117-0fcf-4861-9447-f74de9b3a5e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.797220] env[63355]: DEBUG nova.compute.provider_tree [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.810347] env[63355]: ERROR nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 34286625-39c6-4641-9f1e-99ec49be4177, please check neutron logs for more information. [ 611.810347] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.810347] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.810347] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.810347] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.810347] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.810347] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.810347] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.810347] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.810347] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 611.810347] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.810347] env[63355]: ERROR nova.compute.manager raise self.value [ 611.810347] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.810347] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.810347] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.810347] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.811392] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.811392] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.811392] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 34286625-39c6-4641-9f1e-99ec49be4177, please check neutron logs for more information. [ 611.811392] env[63355]: ERROR nova.compute.manager [ 611.811392] env[63355]: Traceback (most recent call last): [ 611.811392] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.811392] env[63355]: listener.cb(fileno) [ 611.811392] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.811392] env[63355]: result = function(*args, **kwargs) [ 611.811392] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.811392] env[63355]: return func(*args, **kwargs) [ 611.811392] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.811392] env[63355]: raise e [ 611.811392] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.811392] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 611.811392] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.811392] env[63355]: created_port_ids = self._update_ports_for_instance( [ 611.811392] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.811392] env[63355]: with excutils.save_and_reraise_exception(): [ 611.811392] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.811392] env[63355]: self.force_reraise() [ 611.811392] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.811392] env[63355]: raise self.value [ 611.811392] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.811392] env[63355]: updated_port = self._update_port( [ 611.811392] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.811392] env[63355]: _ensure_no_port_binding_failure(port) [ 611.811392] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.811392] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.812907] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 34286625-39c6-4641-9f1e-99ec49be4177, please check neutron logs for more information. [ 611.812907] env[63355]: Removing descriptor: 16 [ 611.812907] env[63355]: ERROR nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 34286625-39c6-4641-9f1e-99ec49be4177, please check neutron logs for more information. [ 611.812907] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Traceback (most recent call last): [ 611.812907] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.812907] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] yield resources [ 611.812907] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.812907] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] self.driver.spawn(context, instance, image_meta, [ 611.812907] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 611.812907] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.812907] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.812907] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] vm_ref = self.build_virtual_machine(instance, [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] for vif in network_info: [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] return self._sync_wrapper(fn, *args, **kwargs) [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] self.wait() [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] self[:] = self._gt.wait() [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] return self._exit_event.wait() [ 611.813485] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] result = hub.switch() [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] return self.greenlet.switch() [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] result = function(*args, **kwargs) [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] return func(*args, **kwargs) [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] raise e [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] nwinfo = self.network_api.allocate_for_instance( [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.814069] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] created_port_ids = self._update_ports_for_instance( [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] with excutils.save_and_reraise_exception(): [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] self.force_reraise() [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] raise self.value [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] updated_port = self._update_port( [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] _ensure_no_port_binding_failure(port) [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.814604] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] raise exception.PortBindingFailed(port_id=port['id']) [ 611.815149] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] nova.exception.PortBindingFailed: Binding failed for port 34286625-39c6-4641-9f1e-99ec49be4177, please check neutron logs for more information. [ 611.815149] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] [ 611.815149] env[63355]: INFO nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Terminating instance [ 611.815149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Acquiring lock "refresh_cache-5c0ba7c6-9946-4765-be4b-d48d3823a68d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.815149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Acquired lock "refresh_cache-5c0ba7c6-9946-4765-be4b-d48d3823a68d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.815149] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.892908] env[63355]: DEBUG nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.012410] env[63355]: DEBUG nova.compute.manager [req-7a2ff941-c683-4b50-877a-d17fd6de1d88 req-a1d47c65-806d-4f31-8652-fd42e2a4d92c service nova] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Received event network-vif-deleted-0423d0dc-4858-4959-91a9-67573bda0806 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.159116] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Successfully created port: a3f82096-5d04-47e8-9c62-dedd206b67ed {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.247635] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 612.250788] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a69dd8c6-f588-4c78-9ca0-8a3d86521971 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.259283] env[63355]: DEBUG oslo_vmware.api [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 612.259283] env[63355]: value = "task-1349473" [ 612.259283] env[63355]: _type = "Task" [ 612.259283] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.275028] env[63355]: DEBUG oslo_vmware.api [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349473, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.301922] env[63355]: DEBUG nova.scheduler.client.report [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.345166] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.481470] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.602223] env[63355]: DEBUG nova.compute.manager [req-26a97957-d114-4d80-8d04-78ed65dd0924 req-a7997b36-8209-4fa4-934d-889fb053a6c9 service nova] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Received event network-changed-34286625-39c6-4641-9f1e-99ec49be4177 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.602223] env[63355]: DEBUG nova.compute.manager [req-26a97957-d114-4d80-8d04-78ed65dd0924 req-a7997b36-8209-4fa4-934d-889fb053a6c9 service nova] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Refreshing instance network info cache due to event network-changed-34286625-39c6-4641-9f1e-99ec49be4177. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 612.602931] env[63355]: DEBUG oslo_concurrency.lockutils [req-26a97957-d114-4d80-8d04-78ed65dd0924 req-a7997b36-8209-4fa4-934d-889fb053a6c9 service nova] Acquiring lock "refresh_cache-5c0ba7c6-9946-4765-be4b-d48d3823a68d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.662870] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquiring lock "abe6726e-7d6a-46e0-9273-de3e681311eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.663191] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lock "abe6726e-7d6a-46e0-9273-de3e681311eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.663443] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquiring lock "abe6726e-7d6a-46e0-9273-de3e681311eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.663617] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lock "abe6726e-7d6a-46e0-9273-de3e681311eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.663819] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lock "abe6726e-7d6a-46e0-9273-de3e681311eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.675464] env[63355]: INFO nova.compute.manager [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Terminating instance [ 612.675464] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquiring lock "refresh_cache-abe6726e-7d6a-46e0-9273-de3e681311eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.675464] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquired lock "refresh_cache-abe6726e-7d6a-46e0-9273-de3e681311eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.675584] env[63355]: DEBUG nova.network.neutron [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.689573] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "2386393a-3006-4e6b-91e0-40b7ab8e3948" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.690189] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "2386393a-3006-4e6b-91e0-40b7ab8e3948" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.772223] env[63355]: DEBUG oslo_vmware.api [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349473, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.811478] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.812570] env[63355]: ERROR nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f0619064-8c90-4914-9df4-a1e6286d3d89, please check neutron logs for more information. [ 612.812570] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Traceback (most recent call last): [ 612.812570] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.812570] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] self.driver.spawn(context, instance, image_meta, [ 612.812570] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 612.812570] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.812570] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.812570] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] vm_ref = self.build_virtual_machine(instance, [ 612.812570] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.812570] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.812570] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] for vif in network_info: [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] return self._sync_wrapper(fn, *args, **kwargs) [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] self.wait() [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] self[:] = self._gt.wait() [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] return self._exit_event.wait() [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] result = hub.switch() [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.814296] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] return self.greenlet.switch() [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] result = function(*args, **kwargs) [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] return func(*args, **kwargs) [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] raise e [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] nwinfo = self.network_api.allocate_for_instance( [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] created_port_ids = self._update_ports_for_instance( [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] with excutils.save_and_reraise_exception(): [ 612.814974] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] self.force_reraise() [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] raise self.value [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] updated_port = self._update_port( [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] _ensure_no_port_binding_failure(port) [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] raise exception.PortBindingFailed(port_id=port['id']) [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] nova.exception.PortBindingFailed: Binding failed for port f0619064-8c90-4914-9df4-a1e6286d3d89, please check neutron logs for more information. [ 612.815426] env[63355]: ERROR nova.compute.manager [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] [ 612.815739] env[63355]: DEBUG nova.compute.utils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Binding failed for port f0619064-8c90-4914-9df4-a1e6286d3d89, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.819034] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.799s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.822761] env[63355]: DEBUG nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Build of instance ce2f828a-82d0-4270-ac21-d967b2e0594b was re-scheduled: Binding failed for port f0619064-8c90-4914-9df4-a1e6286d3d89, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 612.823998] env[63355]: DEBUG nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 612.823998] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "refresh_cache-ce2f828a-82d0-4270-ac21-d967b2e0594b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.823998] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "refresh_cache-ce2f828a-82d0-4270-ac21-d967b2e0594b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.823998] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.906252] env[63355]: DEBUG nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 612.938952] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:24:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2093281057',id=21,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-34736925',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 612.939470] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 612.939794] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 612.940146] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 612.943029] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 612.943029] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 612.943029] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 612.943029] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 612.943029] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 612.943455] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 612.943455] env[63355]: DEBUG nova.virt.hardware [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.943455] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8930bfe8-f02e-42c0-a392-52a76ea12614 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.954676] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1713026d-9258-49d0-9d6b-471823e9c0b7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.985437] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Releasing lock "refresh_cache-5c0ba7c6-9946-4765-be4b-d48d3823a68d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.985437] env[63355]: DEBUG nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.985437] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.985437] env[63355]: DEBUG oslo_concurrency.lockutils [req-26a97957-d114-4d80-8d04-78ed65dd0924 req-a7997b36-8209-4fa4-934d-889fb053a6c9 service nova] Acquired lock "refresh_cache-5c0ba7c6-9946-4765-be4b-d48d3823a68d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.985437] env[63355]: DEBUG nova.network.neutron [req-26a97957-d114-4d80-8d04-78ed65dd0924 req-a7997b36-8209-4fa4-934d-889fb053a6c9 service nova] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Refreshing network info cache for port 34286625-39c6-4641-9f1e-99ec49be4177 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 612.986169] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-323186be-8f70-4cae-a59f-3030e2a82261 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.997835] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a2fa8c-3a75-4895-806d-e3c7f78439d3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.025154] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5c0ba7c6-9946-4765-be4b-d48d3823a68d could not be found. [ 613.025336] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.025514] env[63355]: INFO nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 613.025827] env[63355]: DEBUG oslo.service.loopingcall [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.025973] env[63355]: DEBUG nova.compute.manager [-] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.026083] env[63355]: DEBUG nova.network.neutron [-] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.055224] env[63355]: DEBUG nova.network.neutron [-] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.204688] env[63355]: DEBUG nova.network.neutron [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.269355] env[63355]: DEBUG oslo_vmware.api [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349473, 'name': CreateSnapshot_Task, 'duration_secs': 0.6319} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.269621] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 613.270377] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b37878b-7838-488f-ad91-48983627edd0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.313867] env[63355]: DEBUG nova.network.neutron [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.364016] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.524610] env[63355]: DEBUG nova.network.neutron [req-26a97957-d114-4d80-8d04-78ed65dd0924 req-a7997b36-8209-4fa4-934d-889fb053a6c9 service nova] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.558508] env[63355]: DEBUG nova.network.neutron [-] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.566186] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.634091] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6d2eb4-c8b9-4786-8c6e-8c7356aaaf5d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.643463] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304a8546-7c41-4e55-a648-54794ec791df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.676233] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc405a64-61f9-4845-8ea3-e1f868289d27 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.684667] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674e97ec-5c19-4438-9740-2fcefbfddbd3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.701350] env[63355]: DEBUG nova.compute.provider_tree [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.710229] env[63355]: DEBUG nova.network.neutron [req-26a97957-d114-4d80-8d04-78ed65dd0924 req-a7997b36-8209-4fa4-934d-889fb053a6c9 service nova] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.782435] env[63355]: DEBUG nova.compute.manager [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Instance disappeared during snapshot {{(pid=63355) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 613.816561] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Releasing lock "refresh_cache-abe6726e-7d6a-46e0-9273-de3e681311eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.816965] env[63355]: DEBUG nova.compute.manager [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 613.817245] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 613.819768] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df24e64-8c6e-46fb-a991-11a5e4780fe4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.826893] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 613.826893] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd564db3-26f2-4110-8614-927abc06890f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.834569] env[63355]: DEBUG oslo_vmware.api [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 613.834569] env[63355]: value = "task-1349474" [ 613.834569] env[63355]: _type = "Task" [ 613.834569] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.849626] env[63355]: DEBUG oslo_vmware.api [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349474, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.976020] env[63355]: DEBUG nova.compute.manager [None req-a6ff2a87-fb23-4694-814f-559f8424abd1 tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Found 0 images (rotation: 2) {{(pid=63355) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 614.062013] env[63355]: INFO nova.compute.manager [-] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Took 1.04 seconds to deallocate network for instance. [ 614.064975] env[63355]: DEBUG nova.compute.claims [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.065655] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.069800] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "refresh_cache-ce2f828a-82d0-4270-ac21-d967b2e0594b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.072554] env[63355]: DEBUG nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 614.072554] env[63355]: DEBUG nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.072554] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.094150] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.207198] env[63355]: DEBUG nova.scheduler.client.report [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.214575] env[63355]: DEBUG oslo_concurrency.lockutils [req-26a97957-d114-4d80-8d04-78ed65dd0924 req-a7997b36-8209-4fa4-934d-889fb053a6c9 service nova] Releasing lock "refresh_cache-5c0ba7c6-9946-4765-be4b-d48d3823a68d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.346109] env[63355]: DEBUG oslo_vmware.api [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349474, 'name': PowerOffVM_Task, 'duration_secs': 0.166026} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.346428] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 614.346605] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 614.346861] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f55cd27e-c380-4cbc-b533-c53dc32edfda {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.373404] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 614.373632] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 614.373844] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Deleting the datastore file [datastore2] abe6726e-7d6a-46e0-9273-de3e681311eb {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 614.374131] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90079de6-911e-478f-b1b8-866d78cb06fd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.380978] env[63355]: DEBUG oslo_vmware.api [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for the task: (returnval){ [ 614.380978] env[63355]: value = "task-1349476" [ 614.380978] env[63355]: _type = "Task" [ 614.380978] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.390443] env[63355]: DEBUG oslo_vmware.api [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349476, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.476368] env[63355]: ERROR nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a3f82096-5d04-47e8-9c62-dedd206b67ed, please check neutron logs for more information. [ 614.476368] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.476368] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.476368] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.476368] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.476368] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.476368] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.476368] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.476368] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.476368] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 614.476368] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.476368] env[63355]: ERROR nova.compute.manager raise self.value [ 614.476368] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.476368] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.476368] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.476368] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.476851] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.476851] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.476851] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a3f82096-5d04-47e8-9c62-dedd206b67ed, please check neutron logs for more information. [ 614.476851] env[63355]: ERROR nova.compute.manager [ 614.476851] env[63355]: Traceback (most recent call last): [ 614.476851] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.476851] env[63355]: listener.cb(fileno) [ 614.476851] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.476851] env[63355]: result = function(*args, **kwargs) [ 614.476851] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.476851] env[63355]: return func(*args, **kwargs) [ 614.476851] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.476851] env[63355]: raise e [ 614.476851] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.476851] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 614.476851] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.476851] env[63355]: created_port_ids = self._update_ports_for_instance( [ 614.476851] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.476851] env[63355]: with excutils.save_and_reraise_exception(): [ 614.476851] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.476851] env[63355]: self.force_reraise() [ 614.476851] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.476851] env[63355]: raise self.value [ 614.476851] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.476851] env[63355]: updated_port = self._update_port( [ 614.476851] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.476851] env[63355]: _ensure_no_port_binding_failure(port) [ 614.476851] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.476851] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.477730] env[63355]: nova.exception.PortBindingFailed: Binding failed for port a3f82096-5d04-47e8-9c62-dedd206b67ed, please check neutron logs for more information. [ 614.477730] env[63355]: Removing descriptor: 18 [ 614.477730] env[63355]: ERROR nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a3f82096-5d04-47e8-9c62-dedd206b67ed, please check neutron logs for more information. [ 614.477730] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Traceback (most recent call last): [ 614.477730] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 614.477730] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] yield resources [ 614.477730] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.477730] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] self.driver.spawn(context, instance, image_meta, [ 614.477730] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.477730] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.477730] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.477730] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] vm_ref = self.build_virtual_machine(instance, [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] for vif in network_info: [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] return self._sync_wrapper(fn, *args, **kwargs) [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] self.wait() [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] self[:] = self._gt.wait() [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] return self._exit_event.wait() [ 614.478112] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] result = hub.switch() [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] return self.greenlet.switch() [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] result = function(*args, **kwargs) [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] return func(*args, **kwargs) [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] raise e [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] nwinfo = self.network_api.allocate_for_instance( [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.478438] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] created_port_ids = self._update_ports_for_instance( [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] with excutils.save_and_reraise_exception(): [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] self.force_reraise() [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] raise self.value [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] updated_port = self._update_port( [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] _ensure_no_port_binding_failure(port) [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.478839] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] raise exception.PortBindingFailed(port_id=port['id']) [ 614.479188] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] nova.exception.PortBindingFailed: Binding failed for port a3f82096-5d04-47e8-9c62-dedd206b67ed, please check neutron logs for more information. [ 614.479188] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] [ 614.479188] env[63355]: INFO nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Terminating instance [ 614.480482] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Acquiring lock "refresh_cache-1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.480482] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Acquired lock "refresh_cache-1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.480482] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 614.599474] env[63355]: DEBUG nova.network.neutron [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.711468] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.894s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.712110] env[63355]: ERROR nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 29895cad-8791-4e5d-8afc-ea7cd1027834, please check neutron logs for more information. [ 614.712110] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Traceback (most recent call last): [ 614.712110] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.712110] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] self.driver.spawn(context, instance, image_meta, [ 614.712110] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.712110] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.712110] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.712110] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] vm_ref = self.build_virtual_machine(instance, [ 614.712110] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.712110] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.712110] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] for vif in network_info: [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] return self._sync_wrapper(fn, *args, **kwargs) [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] self.wait() [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] self[:] = self._gt.wait() [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] return self._exit_event.wait() [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] result = hub.switch() [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.712426] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] return self.greenlet.switch() [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] result = function(*args, **kwargs) [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] return func(*args, **kwargs) [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] raise e [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] nwinfo = self.network_api.allocate_for_instance( [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] created_port_ids = self._update_ports_for_instance( [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] with excutils.save_and_reraise_exception(): [ 614.712753] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] self.force_reraise() [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] raise self.value [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] updated_port = self._update_port( [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] _ensure_no_port_binding_failure(port) [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] raise exception.PortBindingFailed(port_id=port['id']) [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] nova.exception.PortBindingFailed: Binding failed for port 29895cad-8791-4e5d-8afc-ea7cd1027834, please check neutron logs for more information. [ 614.713123] env[63355]: ERROR nova.compute.manager [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] [ 614.713407] env[63355]: DEBUG nova.compute.utils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Binding failed for port 29895cad-8791-4e5d-8afc-ea7cd1027834, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 614.717747] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.657s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.717747] env[63355]: INFO nova.compute.claims [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.721611] env[63355]: DEBUG nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Build of instance 8d2ea6a5-5c57-468f-be08-62f418b80e1e was re-scheduled: Binding failed for port 29895cad-8791-4e5d-8afc-ea7cd1027834, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 614.722066] env[63355]: DEBUG nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 614.722296] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Acquiring lock "refresh_cache-8d2ea6a5-5c57-468f-be08-62f418b80e1e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.722442] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Acquired lock "refresh_cache-8d2ea6a5-5c57-468f-be08-62f418b80e1e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.722603] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 614.859047] env[63355]: DEBUG nova.compute.manager [req-64e19c10-4ebd-43ce-baa3-ff5103adc480 req-4d2fe3cc-7ff3-417e-a41d-2599b205f659 service nova] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Received event network-changed-a3f82096-5d04-47e8-9c62-dedd206b67ed {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 614.859161] env[63355]: DEBUG nova.compute.manager [req-64e19c10-4ebd-43ce-baa3-ff5103adc480 req-4d2fe3cc-7ff3-417e-a41d-2599b205f659 service nova] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Refreshing instance network info cache due to event network-changed-a3f82096-5d04-47e8-9c62-dedd206b67ed. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 614.859371] env[63355]: DEBUG oslo_concurrency.lockutils [req-64e19c10-4ebd-43ce-baa3-ff5103adc480 req-4d2fe3cc-7ff3-417e-a41d-2599b205f659 service nova] Acquiring lock "refresh_cache-1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.873518] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquiring lock "115e42ee-60cc-400d-8f4d-c23009f9091f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.873653] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "115e42ee-60cc-400d-8f4d-c23009f9091f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.891105] env[63355]: DEBUG oslo_vmware.api [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Task: {'id': task-1349476, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15448} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.891355] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 614.891555] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 614.891729] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 614.891899] env[63355]: INFO nova.compute.manager [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Took 1.07 seconds to destroy the instance on the hypervisor. [ 614.892271] env[63355]: DEBUG oslo.service.loopingcall [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.895661] env[63355]: DEBUG nova.compute.manager [-] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.895661] env[63355]: DEBUG nova.network.neutron [-] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.915653] env[63355]: DEBUG nova.network.neutron [-] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.012043] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.102580] env[63355]: INFO nova.compute.manager [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] Took 1.03 seconds to deallocate network for instance. [ 615.193165] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.256569] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.412602] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.419549] env[63355]: DEBUG nova.network.neutron [-] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.695280] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Releasing lock "refresh_cache-1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.695887] env[63355]: DEBUG nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 615.695887] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 615.696263] env[63355]: DEBUG oslo_concurrency.lockutils [req-64e19c10-4ebd-43ce-baa3-ff5103adc480 req-4d2fe3cc-7ff3-417e-a41d-2599b205f659 service nova] Acquired lock "refresh_cache-1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.696495] env[63355]: DEBUG nova.network.neutron [req-64e19c10-4ebd-43ce-baa3-ff5103adc480 req-4d2fe3cc-7ff3-417e-a41d-2599b205f659 service nova] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Refreshing network info cache for port a3f82096-5d04-47e8-9c62-dedd206b67ed {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 615.697541] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a1cda36b-6c9e-4d3b-b1a4-edbf9c308281 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.708956] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fabcc9-1fc8-401f-96eb-575918284359 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.734115] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce could not be found. [ 615.735248] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 615.735248] env[63355]: INFO nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Took 0.04 seconds to destroy the instance on the hypervisor. [ 615.735248] env[63355]: DEBUG oslo.service.loopingcall [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.735694] env[63355]: DEBUG nova.compute.manager [-] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.735803] env[63355]: DEBUG nova.network.neutron [-] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.761570] env[63355]: DEBUG nova.network.neutron [-] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.918670] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Releasing lock "refresh_cache-8d2ea6a5-5c57-468f-be08-62f418b80e1e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.919345] env[63355]: DEBUG nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.919904] env[63355]: DEBUG nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.919904] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.926836] env[63355]: INFO nova.compute.manager [-] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Took 1.03 seconds to deallocate network for instance. [ 615.995730] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.066617] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b112a8e3-504a-429c-bb28-2170c1e2f815 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.074782] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Acquiring lock "b46098f3-d952-4a33-86fe-825fe4625008" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.075014] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Lock "b46098f3-d952-4a33-86fe-825fe4625008" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.079463] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce5f81b-4723-4c34-8b2c-bbd8e2c047ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.085221] env[63355]: DEBUG nova.compute.manager [req-030ab459-16f2-49c4-88c4-52967fc1f78a req-d92f6c83-2f87-45ec-aba3-17722ee138b2 service nova] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Received event network-vif-deleted-34286625-39c6-4641-9f1e-99ec49be4177 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.122177] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf788cc-6c2d-4d5e-b5a8-a6b0eea32604 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.133092] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d58227-c076-4439-b843-07f4383a4ee9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.149037] env[63355]: DEBUG nova.compute.provider_tree [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.150933] env[63355]: INFO nova.scheduler.client.report [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleted allocations for instance ce2f828a-82d0-4270-ac21-d967b2e0594b [ 616.221157] env[63355]: DEBUG nova.network.neutron [req-64e19c10-4ebd-43ce-baa3-ff5103adc480 req-4d2fe3cc-7ff3-417e-a41d-2599b205f659 service nova] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.266392] env[63355]: DEBUG nova.network.neutron [-] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.323022] env[63355]: DEBUG nova.network.neutron [req-64e19c10-4ebd-43ce-baa3-ff5103adc480 req-4d2fe3cc-7ff3-417e-a41d-2599b205f659 service nova] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.423033] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "732ff6d3-2de9-4e6b-887d-5329ba83443c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.423033] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "732ff6d3-2de9-4e6b-887d-5329ba83443c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.440641] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.499493] env[63355]: DEBUG nova.network.neutron [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.657698] env[63355]: DEBUG nova.scheduler.client.report [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.660957] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d8dca29a-6394-46a7-871b-427cc7490f88 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "ce2f828a-82d0-4270-ac21-d967b2e0594b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.745s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.662160] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "ce2f828a-82d0-4270-ac21-d967b2e0594b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 20.997s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.662346] env[63355]: INFO nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: ce2f828a-82d0-4270-ac21-d967b2e0594b] During sync_power_state the instance has a pending task (spawning). Skip. [ 616.662516] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "ce2f828a-82d0-4270-ac21-d967b2e0594b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.768280] env[63355]: INFO nova.compute.manager [-] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Took 1.03 seconds to deallocate network for instance. [ 616.770625] env[63355]: DEBUG nova.compute.claims [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 616.770935] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.825710] env[63355]: DEBUG oslo_concurrency.lockutils [req-64e19c10-4ebd-43ce-baa3-ff5103adc480 req-4d2fe3cc-7ff3-417e-a41d-2599b205f659 service nova] Releasing lock "refresh_cache-1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.006802] env[63355]: INFO nova.compute.manager [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] Took 1.08 seconds to deallocate network for instance. [ 617.062279] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "087c0c34-2f44-4791-86e7-b1b3564f49d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.062279] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "087c0c34-2f44-4791-86e7-b1b3564f49d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.097158] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "617b624f-b978-4f98-bbbc-befa006334e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.097158] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "617b624f-b978-4f98-bbbc-befa006334e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.121847] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "12f9b2b4-3c78-43c2-bb9b-d44b81c83123" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.122078] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "12f9b2b4-3c78-43c2-bb9b-d44b81c83123" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.165023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.165023] env[63355]: DEBUG nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 617.166469] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.271s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.168589] env[63355]: INFO nova.compute.claims [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 617.172339] env[63355]: DEBUG nova.compute.manager [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 617.680026] env[63355]: DEBUG nova.compute.utils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.693880] env[63355]: DEBUG nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.693880] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.727172] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.730183] env[63355]: DEBUG nova.compute.manager [req-0e176661-e3cb-4d47-9879-36cad5d3dcd8 req-4b0b5c42-304f-4249-b617-78b0179352cf service nova] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Received event network-vif-deleted-a3f82096-5d04-47e8-9c62-dedd206b67ed {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.827995] env[63355]: DEBUG nova.policy [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '07e7603036e8482fb235dfe6a3538806', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1319b5901cd043029d125bee0a2adb97', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 618.045263] env[63355]: INFO nova.scheduler.client.report [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Deleted allocations for instance 8d2ea6a5-5c57-468f-be08-62f418b80e1e [ 618.188493] env[63355]: DEBUG nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 618.344080] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Successfully created port: b9ffc5f0-96d2-432e-8217-446b9a457ec3 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.555314] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c4f6f6ae-e551-48d7-b47e-fd16276b04b5 tempest-VolumesAssistedSnapshotsTest-183767456 tempest-VolumesAssistedSnapshotsTest-183767456-project-member] Lock "8d2ea6a5-5c57-468f-be08-62f418b80e1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.017s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.556645] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "8d2ea6a5-5c57-468f-be08-62f418b80e1e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 22.891s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.557049] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-822f9fc9-53b4-4037-b0b8-47368fad958a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.567939] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d7a8bd-5e73-4664-a190-5980ec25c38f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.614329] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f0049c-0b42-4f3e-8cdd-ce952b25673c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.624407] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fd6b8f-eb65-4b97-b253-9773b7cc125a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.661135] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7dd8ad-f6ef-413a-b459-007b72c701db {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.669996] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9221fca-d6a2-4b5e-ace0-8dc0152188a0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.689020] env[63355]: DEBUG nova.compute.provider_tree [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.057723] env[63355]: DEBUG nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 619.111179] env[63355]: INFO nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8d2ea6a5-5c57-468f-be08-62f418b80e1e] During the sync_power process the instance has moved from host None to host cpu-1 [ 619.111179] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "8d2ea6a5-5c57-468f-be08-62f418b80e1e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.555s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.191029] env[63355]: DEBUG nova.scheduler.client.report [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.201189] env[63355]: DEBUG nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 619.227903] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.228362] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.228684] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.228984] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.229265] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.229797] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.230133] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.230673] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.230969] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.231294] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.231599] env[63355]: DEBUG nova.virt.hardware [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.232536] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e40ed6-0828-4b66-99de-0ca58db9c88b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.241318] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec1460c-ef72-418c-827e-8860c7b680e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.576373] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquiring lock "ad05eda3-efd7-4040-be9c-6a47df02889b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.577203] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "ad05eda3-efd7-4040-be9c-6a47df02889b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.588717] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.694108] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.528s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.694633] env[63355]: DEBUG nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.697245] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.505s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.698626] env[63355]: INFO nova.compute.claims [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 620.086747] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Acquiring lock "9478ce7a-99c6-4e95-82b0-f8a71ce94a90" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.087106] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Lock "9478ce7a-99c6-4e95-82b0-f8a71ce94a90" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.123493] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "b7dcc653-a430-48c4-9185-3e8200cf1005" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.124270] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "b7dcc653-a430-48c4-9185-3e8200cf1005" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.206978] env[63355]: DEBUG nova.compute.utils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 620.210628] env[63355]: DEBUG nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 620.210820] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 620.265244] env[63355]: DEBUG nova.policy [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4f40faba3c20401383b79aa686b68448', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05714546a8934ccd8b0cdf62d400243b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 620.572633] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Successfully created port: 761cdfb5-da3a-45b9-bba7-b18ba3eac6db {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.720531] env[63355]: DEBUG nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.941315] env[63355]: ERROR nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b9ffc5f0-96d2-432e-8217-446b9a457ec3, please check neutron logs for more information. [ 620.941315] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 620.941315] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.941315] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 620.941315] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.941315] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 620.941315] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.941315] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 620.941315] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.941315] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 620.941315] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.941315] env[63355]: ERROR nova.compute.manager raise self.value [ 620.941315] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.941315] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 620.941315] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.941315] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 620.942036] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.942036] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 620.942036] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b9ffc5f0-96d2-432e-8217-446b9a457ec3, please check neutron logs for more information. [ 620.942036] env[63355]: ERROR nova.compute.manager [ 620.942036] env[63355]: Traceback (most recent call last): [ 620.942036] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 620.942036] env[63355]: listener.cb(fileno) [ 620.942036] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.942036] env[63355]: result = function(*args, **kwargs) [ 620.942036] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.942036] env[63355]: return func(*args, **kwargs) [ 620.942036] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.942036] env[63355]: raise e [ 620.942036] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.942036] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 620.942036] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.942036] env[63355]: created_port_ids = self._update_ports_for_instance( [ 620.942036] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.942036] env[63355]: with excutils.save_and_reraise_exception(): [ 620.942036] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.942036] env[63355]: self.force_reraise() [ 620.942036] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.942036] env[63355]: raise self.value [ 620.942036] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.942036] env[63355]: updated_port = self._update_port( [ 620.942036] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.942036] env[63355]: _ensure_no_port_binding_failure(port) [ 620.942036] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.942036] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 620.942825] env[63355]: nova.exception.PortBindingFailed: Binding failed for port b9ffc5f0-96d2-432e-8217-446b9a457ec3, please check neutron logs for more information. [ 620.942825] env[63355]: Removing descriptor: 18 [ 620.942825] env[63355]: ERROR nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b9ffc5f0-96d2-432e-8217-446b9a457ec3, please check neutron logs for more information. [ 620.942825] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Traceback (most recent call last): [ 620.942825] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 620.942825] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] yield resources [ 620.942825] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.942825] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] self.driver.spawn(context, instance, image_meta, [ 620.942825] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 620.942825] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.942825] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.942825] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] vm_ref = self.build_virtual_machine(instance, [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] for vif in network_info: [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] return self._sync_wrapper(fn, *args, **kwargs) [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] self.wait() [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] self[:] = self._gt.wait() [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] return self._exit_event.wait() [ 620.943737] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] result = hub.switch() [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] return self.greenlet.switch() [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] result = function(*args, **kwargs) [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] return func(*args, **kwargs) [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] raise e [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] nwinfo = self.network_api.allocate_for_instance( [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.944146] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] created_port_ids = self._update_ports_for_instance( [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] with excutils.save_and_reraise_exception(): [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] self.force_reraise() [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] raise self.value [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] updated_port = self._update_port( [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] _ensure_no_port_binding_failure(port) [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.944634] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] raise exception.PortBindingFailed(port_id=port['id']) [ 620.946142] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] nova.exception.PortBindingFailed: Binding failed for port b9ffc5f0-96d2-432e-8217-446b9a457ec3, please check neutron logs for more information. [ 620.946142] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] [ 620.946142] env[63355]: INFO nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Terminating instance [ 620.946777] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Acquiring lock "refresh_cache-c80d746b-ad37-43d7-874e-1c1c848546f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.947666] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Acquired lock "refresh_cache-c80d746b-ad37-43d7-874e-1c1c848546f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.947666] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.219285] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c87fac-eb74-41f2-b452-4a2571deda4f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.234781] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da259440-3e47-422d-a363-6d06236b688f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.266750] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee4a339-5d6f-43dc-bba8-2466a362e6fa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.274592] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf0674b-2311-4e01-af90-9175b0f0c8f6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.288397] env[63355]: DEBUG nova.compute.provider_tree [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.291386] env[63355]: DEBUG nova.compute.manager [req-df29399a-98b8-4e90-8bbf-1821a0710672 req-67fe8159-8248-47dd-9a9b-ef1b6e20c329 service nova] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Received event network-changed-b9ffc5f0-96d2-432e-8217-446b9a457ec3 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 621.291386] env[63355]: DEBUG nova.compute.manager [req-df29399a-98b8-4e90-8bbf-1821a0710672 req-67fe8159-8248-47dd-9a9b-ef1b6e20c329 service nova] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Refreshing instance network info cache due to event network-changed-b9ffc5f0-96d2-432e-8217-446b9a457ec3. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 621.291579] env[63355]: DEBUG oslo_concurrency.lockutils [req-df29399a-98b8-4e90-8bbf-1821a0710672 req-67fe8159-8248-47dd-9a9b-ef1b6e20c329 service nova] Acquiring lock "refresh_cache-c80d746b-ad37-43d7-874e-1c1c848546f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.491023] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.626378] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.737684] env[63355]: DEBUG nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.765397] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.765397] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.765397] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.765564] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.765564] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.765564] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.768201] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.768201] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.768201] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.768201] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.768201] env[63355]: DEBUG nova.virt.hardware [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.769671] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f35e1d-f3a1-4320-a22e-9148c542b30a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.782032] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f6c31c-93cb-4792-a4b0-e2dfcf41b4b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.800994] env[63355]: DEBUG nova.scheduler.client.report [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.044567] env[63355]: DEBUG nova.compute.manager [req-adccf77c-878f-4784-a006-fad2be9bb5ac req-0ff96795-9a3d-405a-a5d2-f5d4a9db7412 service nova] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Received event network-changed-761cdfb5-da3a-45b9-bba7-b18ba3eac6db {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.044567] env[63355]: DEBUG nova.compute.manager [req-adccf77c-878f-4784-a006-fad2be9bb5ac req-0ff96795-9a3d-405a-a5d2-f5d4a9db7412 service nova] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Refreshing instance network info cache due to event network-changed-761cdfb5-da3a-45b9-bba7-b18ba3eac6db. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 622.044567] env[63355]: DEBUG oslo_concurrency.lockutils [req-adccf77c-878f-4784-a006-fad2be9bb5ac req-0ff96795-9a3d-405a-a5d2-f5d4a9db7412 service nova] Acquiring lock "refresh_cache-591a7907-e4a7-41d6-9cd0-5cb359f08810" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.044970] env[63355]: DEBUG oslo_concurrency.lockutils [req-adccf77c-878f-4784-a006-fad2be9bb5ac req-0ff96795-9a3d-405a-a5d2-f5d4a9db7412 service nova] Acquired lock "refresh_cache-591a7907-e4a7-41d6-9cd0-5cb359f08810" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.044970] env[63355]: DEBUG nova.network.neutron [req-adccf77c-878f-4784-a006-fad2be9bb5ac req-0ff96795-9a3d-405a-a5d2-f5d4a9db7412 service nova] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Refreshing network info cache for port 761cdfb5-da3a-45b9-bba7-b18ba3eac6db {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 622.070888] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquiring lock "b2bf4912-6b88-4ece-95c9-e9fd1cf906df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.071182] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lock "b2bf4912-6b88-4ece-95c9-e9fd1cf906df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.084024] env[63355]: ERROR nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 761cdfb5-da3a-45b9-bba7-b18ba3eac6db, please check neutron logs for more information. [ 622.084024] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 622.084024] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.084024] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 622.084024] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.084024] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 622.084024] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.084024] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 622.084024] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.084024] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 622.084024] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.084024] env[63355]: ERROR nova.compute.manager raise self.value [ 622.084024] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.084024] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 622.084024] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.084024] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 622.084783] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.084783] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 622.084783] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 761cdfb5-da3a-45b9-bba7-b18ba3eac6db, please check neutron logs for more information. [ 622.084783] env[63355]: ERROR nova.compute.manager [ 622.084783] env[63355]: Traceback (most recent call last): [ 622.084783] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 622.084783] env[63355]: listener.cb(fileno) [ 622.084783] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.084783] env[63355]: result = function(*args, **kwargs) [ 622.084783] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.084783] env[63355]: return func(*args, **kwargs) [ 622.084783] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.084783] env[63355]: raise e [ 622.084783] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.084783] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 622.084783] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.084783] env[63355]: created_port_ids = self._update_ports_for_instance( [ 622.084783] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.084783] env[63355]: with excutils.save_and_reraise_exception(): [ 622.084783] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.084783] env[63355]: self.force_reraise() [ 622.084783] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.084783] env[63355]: raise self.value [ 622.084783] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.084783] env[63355]: updated_port = self._update_port( [ 622.084783] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.084783] env[63355]: _ensure_no_port_binding_failure(port) [ 622.084783] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.084783] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 622.086223] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 761cdfb5-da3a-45b9-bba7-b18ba3eac6db, please check neutron logs for more information. [ 622.086223] env[63355]: Removing descriptor: 17 [ 622.086223] env[63355]: ERROR nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 761cdfb5-da3a-45b9-bba7-b18ba3eac6db, please check neutron logs for more information. [ 622.086223] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Traceback (most recent call last): [ 622.086223] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 622.086223] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] yield resources [ 622.086223] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.086223] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] self.driver.spawn(context, instance, image_meta, [ 622.086223] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 622.086223] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.086223] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.086223] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] vm_ref = self.build_virtual_machine(instance, [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] for vif in network_info: [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] return self._sync_wrapper(fn, *args, **kwargs) [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] self.wait() [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] self[:] = self._gt.wait() [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] return self._exit_event.wait() [ 622.087594] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] result = hub.switch() [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] return self.greenlet.switch() [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] result = function(*args, **kwargs) [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] return func(*args, **kwargs) [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] raise e [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] nwinfo = self.network_api.allocate_for_instance( [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.087991] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] created_port_ids = self._update_ports_for_instance( [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] with excutils.save_and_reraise_exception(): [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] self.force_reraise() [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] raise self.value [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] updated_port = self._update_port( [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] _ensure_no_port_binding_failure(port) [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.088578] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] raise exception.PortBindingFailed(port_id=port['id']) [ 622.089462] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] nova.exception.PortBindingFailed: Binding failed for port 761cdfb5-da3a-45b9-bba7-b18ba3eac6db, please check neutron logs for more information. [ 622.089462] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] [ 622.089462] env[63355]: INFO nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Terminating instance [ 622.096460] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Acquiring lock "refresh_cache-591a7907-e4a7-41d6-9cd0-5cb359f08810" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.127866] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Releasing lock "refresh_cache-c80d746b-ad37-43d7-874e-1c1c848546f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.128339] env[63355]: DEBUG nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.128528] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.129571] env[63355]: DEBUG oslo_concurrency.lockutils [req-df29399a-98b8-4e90-8bbf-1821a0710672 req-67fe8159-8248-47dd-9a9b-ef1b6e20c329 service nova] Acquired lock "refresh_cache-c80d746b-ad37-43d7-874e-1c1c848546f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.130944] env[63355]: DEBUG nova.network.neutron [req-df29399a-98b8-4e90-8bbf-1821a0710672 req-67fe8159-8248-47dd-9a9b-ef1b6e20c329 service nova] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Refreshing network info cache for port b9ffc5f0-96d2-432e-8217-446b9a457ec3 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 622.131307] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-55863a5a-1a5a-4824-a998-bf25512b46b1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.145617] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fed111-6204-4be3-8ceb-ec5886bb2634 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.175634] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c80d746b-ad37-43d7-874e-1c1c848546f1 could not be found. [ 622.175847] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.176032] env[63355]: INFO nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Took 0.05 seconds to destroy the instance on the hypervisor. [ 622.176320] env[63355]: DEBUG oslo.service.loopingcall [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.176562] env[63355]: DEBUG nova.compute.manager [-] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.176653] env[63355]: DEBUG nova.network.neutron [-] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.198920] env[63355]: DEBUG nova.network.neutron [-] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.306378] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.306911] env[63355]: DEBUG nova.compute.manager [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 622.311819] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.085s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.572494] env[63355]: DEBUG nova.network.neutron [req-adccf77c-878f-4784-a006-fad2be9bb5ac req-0ff96795-9a3d-405a-a5d2-f5d4a9db7412 service nova] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.679406] env[63355]: DEBUG nova.network.neutron [req-df29399a-98b8-4e90-8bbf-1821a0710672 req-67fe8159-8248-47dd-9a9b-ef1b6e20c329 service nova] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.682067] env[63355]: DEBUG nova.network.neutron [req-adccf77c-878f-4784-a006-fad2be9bb5ac req-0ff96795-9a3d-405a-a5d2-f5d4a9db7412 service nova] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.704092] env[63355]: DEBUG nova.network.neutron [-] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.814086] env[63355]: DEBUG nova.compute.utils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.816047] env[63355]: DEBUG nova.compute.manager [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Not allocating networking since 'none' was specified. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 622.822223] env[63355]: DEBUG nova.network.neutron [req-df29399a-98b8-4e90-8bbf-1821a0710672 req-67fe8159-8248-47dd-9a9b-ef1b6e20c329 service nova] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.185780] env[63355]: DEBUG oslo_concurrency.lockutils [req-adccf77c-878f-4784-a006-fad2be9bb5ac req-0ff96795-9a3d-405a-a5d2-f5d4a9db7412 service nova] Releasing lock "refresh_cache-591a7907-e4a7-41d6-9cd0-5cb359f08810" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.186418] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Acquired lock "refresh_cache-591a7907-e4a7-41d6-9cd0-5cb359f08810" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.186604] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 623.208424] env[63355]: INFO nova.compute.manager [-] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Took 1.03 seconds to deallocate network for instance. [ 623.209373] env[63355]: DEBUG nova.compute.claims [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.209553] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.214601] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c5c2b8-09b6-4650-b4bb-531a3ad07ce9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.222190] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579bfe9e-c554-4209-b188-bc8ab880f002 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.256874] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f6f219-1a97-45db-918b-68eba2635f45 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.264666] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1b9646-f544-4fc6-9dda-acedbae4ea22 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.278806] env[63355]: DEBUG nova.compute.provider_tree [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.319015] env[63355]: DEBUG nova.compute.manager [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 623.326686] env[63355]: DEBUG oslo_concurrency.lockutils [req-df29399a-98b8-4e90-8bbf-1821a0710672 req-67fe8159-8248-47dd-9a9b-ef1b6e20c329 service nova] Releasing lock "refresh_cache-c80d746b-ad37-43d7-874e-1c1c848546f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.687603] env[63355]: DEBUG nova.compute.manager [req-c039bdb9-b228-4662-b4ed-736cc172d968 req-c1042f93-97eb-4e0f-8429-a6a9f8e5f9e9 service nova] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Received event network-vif-deleted-b9ffc5f0-96d2-432e-8217-446b9a457ec3 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 623.718622] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.781812] env[63355]: DEBUG nova.scheduler.client.report [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.965652] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.293406] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.293406] env[63355]: ERROR nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0423d0dc-4858-4959-91a9-67573bda0806, please check neutron logs for more information. [ 624.293406] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Traceback (most recent call last): [ 624.293406] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.293406] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] self.driver.spawn(context, instance, image_meta, [ 624.293406] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 624.293406] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.293406] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.293406] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] vm_ref = self.build_virtual_machine(instance, [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] for vif in network_info: [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] return self._sync_wrapper(fn, *args, **kwargs) [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] self.wait() [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] self[:] = self._gt.wait() [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] return self._exit_event.wait() [ 624.294142] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] result = hub.switch() [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] return self.greenlet.switch() [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] result = function(*args, **kwargs) [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] return func(*args, **kwargs) [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] raise e [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] nwinfo = self.network_api.allocate_for_instance( [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.294517] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] created_port_ids = self._update_ports_for_instance( [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] with excutils.save_and_reraise_exception(): [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] self.force_reraise() [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] raise self.value [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] updated_port = self._update_port( [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] _ensure_no_port_binding_failure(port) [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.294862] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] raise exception.PortBindingFailed(port_id=port['id']) [ 624.295203] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] nova.exception.PortBindingFailed: Binding failed for port 0423d0dc-4858-4959-91a9-67573bda0806, please check neutron logs for more information. [ 624.295203] env[63355]: ERROR nova.compute.manager [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] [ 624.295203] env[63355]: DEBUG nova.compute.utils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Binding failed for port 0423d0dc-4858-4959-91a9-67573bda0806, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.295203] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.802s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.299745] env[63355]: INFO nova.compute.claims [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.303921] env[63355]: DEBUG nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Build of instance 0fb6ba21-d45e-4809-8e68-d78a5892baeb was re-scheduled: Binding failed for port 0423d0dc-4858-4959-91a9-67573bda0806, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 624.305787] env[63355]: DEBUG nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 624.305787] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Acquiring lock "refresh_cache-0fb6ba21-d45e-4809-8e68-d78a5892baeb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.305787] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Acquired lock "refresh_cache-0fb6ba21-d45e-4809-8e68-d78a5892baeb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.305787] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.331532] env[63355]: DEBUG nova.compute.manager [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 624.352028] env[63355]: DEBUG nova.compute.manager [req-96e6d069-811a-400c-9515-822dd05e5fda req-a31337b7-f784-40f9-9552-10f738b20275 service nova] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Received event network-vif-deleted-761cdfb5-da3a-45b9-bba7-b18ba3eac6db {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.365482] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 624.365751] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 624.365903] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.366310] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 624.366460] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.366601] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 624.366876] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 624.367109] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 624.367280] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 624.367557] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.367715] env[63355]: DEBUG nova.virt.hardware [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.369030] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8afba2-0268-40ca-9819-ba0f6b5910d6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.378736] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde2a986-2728-44e6-91c1-05266fe09690 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.397982] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 624.407632] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Creating folder: Project (7a05ed71e530468fb884edbba34e4836). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 624.407632] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-517561cb-cf44-48ce-8535-d34cb78addb7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.416374] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Created folder: Project (7a05ed71e530468fb884edbba34e4836) in parent group-v287607. [ 624.416569] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Creating folder: Instances. Parent ref: group-v287612. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 624.416804] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b5b1049-322d-4f74-a6a1-f10e8182e5ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.425720] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Created folder: Instances in parent group-v287612. [ 624.425958] env[63355]: DEBUG oslo.service.loopingcall [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.426169] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 624.426370] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d43f3e79-3a8a-41e0-b488-0ce11684633f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.450227] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 624.450227] env[63355]: value = "task-1349479" [ 624.450227] env[63355]: _type = "Task" [ 624.450227] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.460897] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349479, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.472405] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Releasing lock "refresh_cache-591a7907-e4a7-41d6-9cd0-5cb359f08810" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.472834] env[63355]: DEBUG nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 624.473090] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 624.473536] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e9fb888-9260-43d9-a458-ee6d5e9bcf3d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.482176] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6272c477-a800-4263-9ef8-733627adbbaa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.504671] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 591a7907-e4a7-41d6-9cd0-5cb359f08810 could not be found. [ 624.504784] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 624.504935] env[63355]: INFO nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Took 0.03 seconds to destroy the instance on the hypervisor. [ 624.505612] env[63355]: DEBUG oslo.service.loopingcall [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.505743] env[63355]: DEBUG nova.compute.manager [-] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.505833] env[63355]: DEBUG nova.network.neutron [-] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 624.610351] env[63355]: DEBUG nova.network.neutron [-] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.853236] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.961789] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349479, 'name': CreateVM_Task, 'duration_secs': 0.25829} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.966250] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 624.971683] env[63355]: DEBUG oslo_vmware.service [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04c0baa-820c-45ce-83bf-b46df84bb271 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.974374] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "e5c80c05-0e48-415a-ac91-a53e1c8819f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.974818] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "e5c80c05-0e48-415a-ac91-a53e1c8819f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.979431] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.982043] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.982043] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 624.982043] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e562903a-0ca7-463f-acd6-7d5e4372addc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.986918] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 624.986918] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527d3842-2e7c-f0d5-f45e-f69778181774" [ 624.986918] env[63355]: _type = "Task" [ 624.986918] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.997366] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527d3842-2e7c-f0d5-f45e-f69778181774, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.064283] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.113314] env[63355]: DEBUG nova.network.neutron [-] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.507023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.507968] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 625.507968] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.508121] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.508399] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 625.511556] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76e8cf9a-c9e9-40ae-aef6-26c544bf2e97 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.569224] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Releasing lock "refresh_cache-0fb6ba21-d45e-4809-8e68-d78a5892baeb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.569471] env[63355]: DEBUG nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.569650] env[63355]: DEBUG nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.569816] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.615237] env[63355]: INFO nova.compute.manager [-] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Took 1.11 seconds to deallocate network for instance. [ 625.621248] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.622447] env[63355]: DEBUG nova.compute.claims [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 625.622614] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.815772] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ac56fc-c4a6-4b19-9f1e-66885dac2dcd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.823373] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce1da96-bbec-4579-980d-d4213a5727a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.857752] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7110f1-a749-43cd-925c-0ea402d3f91d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.865371] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44b61d6-d3d3-4739-bcea-92846684473f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.878845] env[63355]: DEBUG nova.compute.provider_tree [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.125609] env[63355]: DEBUG nova.network.neutron [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.384901] env[63355]: DEBUG nova.scheduler.client.report [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.628961] env[63355]: INFO nova.compute.manager [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] [instance: 0fb6ba21-d45e-4809-8e68-d78a5892baeb] Took 1.06 seconds to deallocate network for instance. [ 626.870858] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 626.871268] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 626.874603] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387ba40f-77b3-44ea-b373-0e3ae8d96e46 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.879182] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2a04b6e-b47a-448b-b723-1d56a5b9352b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.890137] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.597s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.890640] env[63355]: DEBUG nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 626.895017] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 626.895017] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526940c3-f895-388e-b167-b04bcb5e44a2" [ 626.895017] env[63355]: _type = "Task" [ 626.895017] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.902019] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.908s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.915667] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Preparing fetch location {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 626.915667] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Creating directory with path [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 626.915667] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f58ac977-e5a8-44ff-86f0-3a9824772e74 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.936361] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Created directory with path [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 626.936361] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Fetch image to [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 626.936361] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Downloading image file data 84ca0bb3-9916-4cea-9399-75af77b0558e to [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk on the data store datastore1 {{(pid=63355) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 626.936361] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1ea10a-a3d6-411c-a21b-155e1713eb3c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.962067] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e648b8-cdab-43a4-869d-b91b88666034 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.979430] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13fedebf-699f-489d-8b6a-8ef178864228 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.014022] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ec260a-37b0-4c40-a4ab-be0c169e5558 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.027233] env[63355]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-237f8fc6-9a2b-4113-806d-8359bfd3f0c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.047608] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Downloading image file data 84ca0bb3-9916-4cea-9399-75af77b0558e to the data store datastore1 {{(pid=63355) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 627.143668] env[63355]: DEBUG oslo_vmware.rw_handles [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63355) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 627.404789] env[63355]: DEBUG nova.compute.utils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.407498] env[63355]: DEBUG nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 627.407596] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 627.526453] env[63355]: DEBUG nova.policy [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a3f5c5e193844989a127f2ef3c5a2521', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c852f4f675eb4836a8169e272f1c2371', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 627.682310] env[63355]: INFO nova.scheduler.client.report [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Deleted allocations for instance 0fb6ba21-d45e-4809-8e68-d78a5892baeb [ 627.847304] env[63355]: DEBUG oslo_vmware.rw_handles [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Completed reading data from the image iterator. {{(pid=63355) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 627.847794] env[63355]: DEBUG oslo_vmware.rw_handles [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 627.915934] env[63355]: DEBUG nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 627.951776] env[63355]: WARNING nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance abe6726e-7d6a-46e0-9273-de3e681311eb is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 627.951931] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 5c0ba7c6-9946-4765-be4b-d48d3823a68d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 627.952074] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 627.952193] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance c80d746b-ad37-43d7-874e-1c1c848546f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 627.952306] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 591a7907-e4a7-41d6-9cd0-5cb359f08810 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 627.952414] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance def90353-ae90-4e1e-9fe1-8e5459a7309d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 627.952520] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 4392313f-2fed-4308-b79a-324b3431d5af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 627.982920] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Downloaded image file data 84ca0bb3-9916-4cea-9399-75af77b0558e to vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk on the data store datastore1 {{(pid=63355) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 627.984959] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Caching image {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 627.985484] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Copying Virtual Disk [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk to [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 627.985780] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94e1fd7a-bf12-4a3a-ab5c-ebdd593951d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.000798] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 628.000798] env[63355]: value = "task-1349480" [ 628.000798] env[63355]: _type = "Task" [ 628.000798] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.011816] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349480, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.196216] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2221afc8-6b2b-40af-b495-57d814d70a89 tempest-ImagesNegativeTestJSON-1699209473 tempest-ImagesNegativeTestJSON-1699209473-project-member] Lock "0fb6ba21-d45e-4809-8e68-d78a5892baeb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.205s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.457164] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance bc1d7402-3651-488b-a431-2cb2ae987d32 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 628.514836] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349480, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.575178] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Successfully created port: b60350f9-d151-4fea-8aad-7d681d5936b9 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.699582] env[63355]: DEBUG nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.929686] env[63355]: DEBUG nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 628.961332] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance f9c8ed86-506b-4654-8e7e-cb218605764b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 628.969298] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.969298] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.969298] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.969298] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.969577] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.969577] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.969577] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.969577] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.969577] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.969808] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.969808] env[63355]: DEBUG nova.virt.hardware [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.970945] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79826473-b2c3-4495-8445-9daacf0c558c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.983134] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7151c36d-b42b-4a03-a623-73085f090f02 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.011633] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349480, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.226510] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.476279] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 1de0e894-1edb-4296-9e9b-2c1eb0c7c275 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 629.514412] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349480, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.979979] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 10112b62-f41e-4413-9bc8-c1306c5c1bc1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 630.015460] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349480, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.668559} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.016813] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Copied Virtual Disk [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk to [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 630.017011] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleting the datastore file [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e/tmp-sparse.vmdk {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 630.017283] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-712c87a3-0c82-4832-aec6-f82128c550a4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.023901] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 630.023901] env[63355]: value = "task-1349481" [ 630.023901] env[63355]: _type = "Task" [ 630.023901] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.031671] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349481, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.484377] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b275344f-695e-463b-8bfe-2bee97296349 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 630.534368] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349481, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023381} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.534625] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 630.534840] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Moving file from [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595/84ca0bb3-9916-4cea-9399-75af77b0558e to [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e. {{(pid=63355) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 630.535367] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-3ff53838-41ae-4be4-a444-2ff662ce0180 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.542713] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 630.542713] env[63355]: value = "task-1349486" [ 630.542713] env[63355]: _type = "Task" [ 630.542713] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.550815] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349486, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.624920] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "9dd8b658-886a-4f11-ab9b-059c4265436c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.625364] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "9dd8b658-886a-4f11-ab9b-059c4265436c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.656272] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Acquiring lock "d7589681-fe6c-4dc6-b533-ade8ae634d58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.656529] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Lock "d7589681-fe6c-4dc6-b533-ade8ae634d58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.988744] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 631.054940] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349486, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.027522} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.054940] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] File moved {{(pid=63355) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 631.054940] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Cleaning up location [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595 {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 631.054940] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleting the datastore file [datastore1] vmware_temp/400b1641-0eb5-4f53-805a-ec6445b7a595 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 631.055340] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-000c681c-187e-4b7c-991d-880dca298da4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.062175] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 631.062175] env[63355]: value = "task-1349487" [ 631.062175] env[63355]: _type = "Task" [ 631.062175] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.070780] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349487, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.105842] env[63355]: ERROR nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b60350f9-d151-4fea-8aad-7d681d5936b9, please check neutron logs for more information. [ 631.105842] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.105842] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.105842] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.105842] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.105842] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.105842] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.105842] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.105842] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.105842] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 631.105842] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.105842] env[63355]: ERROR nova.compute.manager raise self.value [ 631.105842] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.105842] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.105842] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.105842] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.106430] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.106430] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.106430] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b60350f9-d151-4fea-8aad-7d681d5936b9, please check neutron logs for more information. [ 631.106430] env[63355]: ERROR nova.compute.manager [ 631.106430] env[63355]: Traceback (most recent call last): [ 631.106430] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.106430] env[63355]: listener.cb(fileno) [ 631.106430] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.106430] env[63355]: result = function(*args, **kwargs) [ 631.106430] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.106430] env[63355]: return func(*args, **kwargs) [ 631.106430] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.106430] env[63355]: raise e [ 631.106430] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.106430] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 631.106430] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.106430] env[63355]: created_port_ids = self._update_ports_for_instance( [ 631.106430] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.106430] env[63355]: with excutils.save_and_reraise_exception(): [ 631.106430] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.106430] env[63355]: self.force_reraise() [ 631.106430] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.106430] env[63355]: raise self.value [ 631.106430] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.106430] env[63355]: updated_port = self._update_port( [ 631.106430] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.106430] env[63355]: _ensure_no_port_binding_failure(port) [ 631.106430] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.106430] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.108558] env[63355]: nova.exception.PortBindingFailed: Binding failed for port b60350f9-d151-4fea-8aad-7d681d5936b9, please check neutron logs for more information. [ 631.108558] env[63355]: Removing descriptor: 18 [ 631.108558] env[63355]: ERROR nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b60350f9-d151-4fea-8aad-7d681d5936b9, please check neutron logs for more information. [ 631.108558] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Traceback (most recent call last): [ 631.108558] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.108558] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] yield resources [ 631.108558] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.108558] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] self.driver.spawn(context, instance, image_meta, [ 631.108558] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 631.108558] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.108558] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.108558] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] vm_ref = self.build_virtual_machine(instance, [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] for vif in network_info: [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] return self._sync_wrapper(fn, *args, **kwargs) [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] self.wait() [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] self[:] = self._gt.wait() [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] return self._exit_event.wait() [ 631.109290] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] result = hub.switch() [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] return self.greenlet.switch() [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] result = function(*args, **kwargs) [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] return func(*args, **kwargs) [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] raise e [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] nwinfo = self.network_api.allocate_for_instance( [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.109738] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] created_port_ids = self._update_ports_for_instance( [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] with excutils.save_and_reraise_exception(): [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] self.force_reraise() [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] raise self.value [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] updated_port = self._update_port( [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] _ensure_no_port_binding_failure(port) [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.110152] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] raise exception.PortBindingFailed(port_id=port['id']) [ 631.110644] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] nova.exception.PortBindingFailed: Binding failed for port b60350f9-d151-4fea-8aad-7d681d5936b9, please check neutron logs for more information. [ 631.110644] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] [ 631.110644] env[63355]: INFO nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Terminating instance [ 631.110644] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Acquiring lock "refresh_cache-4392313f-2fed-4308-b79a-324b3431d5af" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.110644] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Acquired lock "refresh_cache-4392313f-2fed-4308-b79a-324b3431d5af" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.110644] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.186814] env[63355]: DEBUG nova.compute.manager [req-1f222274-9088-4484-ac8a-47e69d699c3a req-5d7234c3-02b3-4ecd-8034-c74ca36605fd service nova] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Received event network-changed-b60350f9-d151-4fea-8aad-7d681d5936b9 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.188303] env[63355]: DEBUG nova.compute.manager [req-1f222274-9088-4484-ac8a-47e69d699c3a req-5d7234c3-02b3-4ecd-8034-c74ca36605fd service nova] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Refreshing instance network info cache due to event network-changed-b60350f9-d151-4fea-8aad-7d681d5936b9. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 631.189706] env[63355]: DEBUG oslo_concurrency.lockutils [req-1f222274-9088-4484-ac8a-47e69d699c3a req-5d7234c3-02b3-4ecd-8034-c74ca36605fd service nova] Acquiring lock "refresh_cache-4392313f-2fed-4308-b79a-324b3431d5af" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.493614] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 0de6892a-5463-4d60-807c-0e3fca2accf3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 631.580545] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "17a904fe-ab4b-46ed-98c8-175987710643" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.582888] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "17a904fe-ab4b-46ed-98c8-175987710643" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.582888] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349487, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16743} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.582888] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 631.587019] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22b6f8a5-e07d-44ad-8ff8-2c9dbf2ad1ca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.594567] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 631.594567] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52234009-6676-730c-da80-bf2a88ad7704" [ 631.594567] env[63355]: _type = "Task" [ 631.594567] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.605201] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52234009-6676-730c-da80-bf2a88ad7704, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.825631] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.997242] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 2386393a-3006-4e6b-91e0-40b7ab8e3948 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 632.010747] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.107565] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52234009-6676-730c-da80-bf2a88ad7704, 'name': SearchDatastore_Task, 'duration_secs': 0.011421} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.107776] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.109304] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] def90353-ae90-4e1e-9fe1-8e5459a7309d/def90353-ae90-4e1e-9fe1-8e5459a7309d.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 632.109304] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ac8a9ef-8249-4eee-8bde-6fbb8f5670f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.116657] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 632.116657] env[63355]: value = "task-1349488" [ 632.116657] env[63355]: _type = "Task" [ 632.116657] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.127726] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349488, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.377815] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.378146] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.500777] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 115e42ee-60cc-400d-8f4d-c23009f9091f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 632.513661] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Releasing lock "refresh_cache-4392313f-2fed-4308-b79a-324b3431d5af" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.514129] env[63355]: DEBUG nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.517331] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.517789] env[63355]: DEBUG oslo_concurrency.lockutils [req-1f222274-9088-4484-ac8a-47e69d699c3a req-5d7234c3-02b3-4ecd-8034-c74ca36605fd service nova] Acquired lock "refresh_cache-4392313f-2fed-4308-b79a-324b3431d5af" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.517976] env[63355]: DEBUG nova.network.neutron [req-1f222274-9088-4484-ac8a-47e69d699c3a req-5d7234c3-02b3-4ecd-8034-c74ca36605fd service nova] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Refreshing network info cache for port b60350f9-d151-4fea-8aad-7d681d5936b9 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 632.520462] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7211512-8ddc-44d3-8952-8b409d86ac77 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.530762] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4bfa596-810d-493c-9a41-ce3cab93f475 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.557374] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4392313f-2fed-4308-b79a-324b3431d5af could not be found. [ 632.557495] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.558174] env[63355]: INFO nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Took 0.04 seconds to destroy the instance on the hypervisor. [ 632.558895] env[63355]: DEBUG oslo.service.loopingcall [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.558895] env[63355]: DEBUG nova.compute.manager [-] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.559071] env[63355]: DEBUG nova.network.neutron [-] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.612395] env[63355]: DEBUG nova.network.neutron [-] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.635241] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349488, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.009156] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b46098f3-d952-4a33-86fe-825fe4625008 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 633.050898] env[63355]: DEBUG nova.network.neutron [req-1f222274-9088-4484-ac8a-47e69d699c3a req-5d7234c3-02b3-4ecd-8034-c74ca36605fd service nova] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.115775] env[63355]: DEBUG nova.network.neutron [-] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.136274] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349488, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533954} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.136608] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] def90353-ae90-4e1e-9fe1-8e5459a7309d/def90353-ae90-4e1e-9fe1-8e5459a7309d.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 633.136757] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 633.140019] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84bc3d45-d231-47e0-bb78-6efb92e7ca30 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.147986] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 633.147986] env[63355]: value = "task-1349490" [ 633.147986] env[63355]: _type = "Task" [ 633.147986] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.158439] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349490, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.218833] env[63355]: DEBUG nova.network.neutron [req-1f222274-9088-4484-ac8a-47e69d699c3a req-5d7234c3-02b3-4ecd-8034-c74ca36605fd service nova] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.418623] env[63355]: DEBUG nova.compute.manager [req-275d4378-e671-44a9-8335-c6d5f1f05d26 req-32b5de0e-3bd8-44a1-aa78-897a6b49aeb3 service nova] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Received event network-vif-deleted-b60350f9-d151-4fea-8aad-7d681d5936b9 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.518728] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 732ff6d3-2de9-4e6b-887d-5329ba83443c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 633.619788] env[63355]: INFO nova.compute.manager [-] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Took 1.06 seconds to deallocate network for instance. [ 633.623125] env[63355]: DEBUG nova.compute.claims [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 633.623298] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.658287] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349490, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071625} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.658626] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 633.659640] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131b2269-aeb0-465f-bf27-02a562bb5382 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.689682] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Reconfiguring VM instance instance-0000000e to attach disk [datastore1] def90353-ae90-4e1e-9fe1-8e5459a7309d/def90353-ae90-4e1e-9fe1-8e5459a7309d.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 633.692282] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0131c6f-5bb2-4d39-81d6-fa36e50547c5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.715784] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 633.715784] env[63355]: value = "task-1349491" [ 633.715784] env[63355]: _type = "Task" [ 633.715784] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.724742] env[63355]: DEBUG oslo_concurrency.lockutils [req-1f222274-9088-4484-ac8a-47e69d699c3a req-5d7234c3-02b3-4ecd-8034-c74ca36605fd service nova] Releasing lock "refresh_cache-4392313f-2fed-4308-b79a-324b3431d5af" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.725405] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349491, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.025535] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 087c0c34-2f44-4791-86e7-b1b3564f49d5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 634.227562] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349491, 'name': ReconfigVM_Task, 'duration_secs': 0.293208} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.228154] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Reconfigured VM instance instance-0000000e to attach disk [datastore1] def90353-ae90-4e1e-9fe1-8e5459a7309d/def90353-ae90-4e1e-9fe1-8e5459a7309d.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 634.228931] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2e52ce5-66b9-471e-8a62-496e9a3d482b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.237652] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 634.237652] env[63355]: value = "task-1349492" [ 634.237652] env[63355]: _type = "Task" [ 634.237652] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.246358] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349492, 'name': Rename_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.444019] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Acquiring lock "8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.444380] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Lock "8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.529834] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 617b624f-b978-4f98-bbbc-befa006334e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 634.747809] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349492, 'name': Rename_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.034157] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 12f9b2b4-3c78-43c2-bb9b-d44b81c83123 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.246133] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349492, 'name': Rename_Task, 'duration_secs': 0.841813} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.246432] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 635.246674] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31a5ba1e-584b-4132-ba69-25837fd59c99 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.257058] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 635.257058] env[63355]: value = "task-1349494" [ 635.257058] env[63355]: _type = "Task" [ 635.257058] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.266823] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349494, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.537359] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance ad05eda3-efd7-4040-be9c-6a47df02889b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.766794] env[63355]: DEBUG oslo_vmware.api [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349494, 'name': PowerOnVM_Task, 'duration_secs': 0.469813} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.767395] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 635.767665] env[63355]: INFO nova.compute.manager [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Took 11.44 seconds to spawn the instance on the hypervisor. [ 635.768107] env[63355]: DEBUG nova.compute.manager [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 635.768978] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d7db39-c55f-4cef-86a1-ee170f193c02 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.040737] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 9478ce7a-99c6-4e95-82b0-f8a71ce94a90 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 636.295561] env[63355]: INFO nova.compute.manager [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Took 29.13 seconds to build instance. [ 636.545858] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b7dcc653-a430-48c4-9185-3e8200cf1005 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 636.801457] env[63355]: DEBUG oslo_concurrency.lockutils [None req-92cbcdb5-96b8-4abe-95ab-9516f10dee3e tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "def90353-ae90-4e1e-9fe1-8e5459a7309d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.842s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.050620] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b2bf4912-6b88-4ece-95c9-e9fd1cf906df has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.305018] env[63355]: DEBUG nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.554811] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance e5c80c05-0e48-415a-ac91-a53e1c8819f0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.554891] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 637.555233] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 637.836741] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.054148] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c74e4f4-90d4-4d58-a159-e5e63edadd03 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.061912] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5951a81e-7d10-4c45-b526-ac770729d74d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.098920] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4eb92de-a669-4bbb-85f4-a76448058022 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.106277] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6269036a-e346-457e-9e99-73d445d5a565 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.119677] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.623837] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.128692] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 639.128974] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.234s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.130178] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.114s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.130874] env[63355]: INFO nova.compute.claims [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 639.140124] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 639.140124] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Getting list of instances from cluster (obj){ [ 639.140124] env[63355]: value = "domain-c8" [ 639.140124] env[63355]: _type = "ClusterComputeResource" [ 639.140124] env[63355]: } {{(pid=63355) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 639.141233] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bfdc074-ce6e-477f-a9cc-f1b3753b8c2e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.157387] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Got total of 1 instances {{(pid=63355) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 639.224396] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "1446c7da-415f-43ee-9ff1-2266bd3038e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.224396] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "1446c7da-415f-43ee-9ff1-2266bd3038e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.571835] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Acquiring lock "ec4726b2-38f9-4499-a5b0-a70db0d697f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.572125] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Lock "ec4726b2-38f9-4499-a5b0-a70db0d697f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.613321] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d161424-9336-4652-9b28-79e915199b70 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.621350] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f9b35b-c590-48c9-b95c-31228b967106 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.658053] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d96bcc-3782-4ef8-8080-1463c2d6a1f7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.665972] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6434348-222e-4d65-9fb2-ef25bd320aef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.679134] env[63355]: DEBUG nova.compute.provider_tree [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.182412] env[63355]: DEBUG nova.scheduler.client.report [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.687962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.688531] env[63355]: DEBUG nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 641.695660] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.626s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.197125] env[63355]: DEBUG nova.compute.utils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 642.203115] env[63355]: DEBUG nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 642.203115] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 642.271611] env[63355]: DEBUG nova.policy [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1762b48bb9a24028a12b2ac455353f04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7274335fcc1441e3b858591ac415aec4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 642.677851] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea51023-116c-44a1-8f9c-7656e7a69088 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.686139] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308b86fe-1f27-4d77-b16b-f92f977eaa74 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.723145] env[63355]: DEBUG nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 642.725444] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11544067-08dc-4499-af3d-d23d830f2c8b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.735204] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08bd7114-0c49-4b0d-af6e-7b158957a2bd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.749197] env[63355]: DEBUG nova.compute.provider_tree [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.794516] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Successfully created port: 6b8ca815-f009-4f5f-9095-66890ffdef10 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 642.963252] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Acquiring lock "3c04c546-a9fa-4cbd-98d4-3ea92918a918" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.963496] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Lock "3c04c546-a9fa-4cbd-98d4-3ea92918a918" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.257084] env[63355]: DEBUG nova.scheduler.client.report [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.736319] env[63355]: DEBUG nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 643.759936] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.068s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.760486] env[63355]: ERROR nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 34286625-39c6-4641-9f1e-99ec49be4177, please check neutron logs for more information. [ 643.760486] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Traceback (most recent call last): [ 643.760486] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 643.760486] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] self.driver.spawn(context, instance, image_meta, [ 643.760486] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 643.760486] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.760486] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.760486] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] vm_ref = self.build_virtual_machine(instance, [ 643.760486] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.760486] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.760486] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] for vif in network_info: [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] return self._sync_wrapper(fn, *args, **kwargs) [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] self.wait() [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] self[:] = self._gt.wait() [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] return self._exit_event.wait() [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] result = hub.switch() [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 643.760949] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] return self.greenlet.switch() [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] result = function(*args, **kwargs) [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] return func(*args, **kwargs) [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] raise e [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] nwinfo = self.network_api.allocate_for_instance( [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] created_port_ids = self._update_ports_for_instance( [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] with excutils.save_and_reraise_exception(): [ 643.761412] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] self.force_reraise() [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] raise self.value [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] updated_port = self._update_port( [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] _ensure_no_port_binding_failure(port) [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] raise exception.PortBindingFailed(port_id=port['id']) [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] nova.exception.PortBindingFailed: Binding failed for port 34286625-39c6-4641-9f1e-99ec49be4177, please check neutron logs for more information. [ 643.761877] env[63355]: ERROR nova.compute.manager [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] [ 643.762290] env[63355]: DEBUG nova.compute.utils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Binding failed for port 34286625-39c6-4641-9f1e-99ec49be4177, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 643.767591] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 643.767810] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 643.767966] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 643.768160] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 643.768307] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 643.768490] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 643.768630] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 643.769019] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 643.769019] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 643.769107] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 643.769251] env[63355]: DEBUG nova.virt.hardware [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 643.769557] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.330s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.769730] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.771611] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.774897] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab36c98-196a-4b5b-bc76-ddf154778243 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.778031] env[63355]: DEBUG nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Build of instance 5c0ba7c6-9946-4765-be4b-d48d3823a68d was re-scheduled: Binding failed for port 34286625-39c6-4641-9f1e-99ec49be4177, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 643.778901] env[63355]: DEBUG nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 643.778901] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Acquiring lock "refresh_cache-5c0ba7c6-9946-4765-be4b-d48d3823a68d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.778901] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Acquired lock "refresh_cache-5c0ba7c6-9946-4765-be4b-d48d3823a68d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.779062] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.785845] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b095f00a-4055-463f-96ba-bc89b407df90 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.803919] env[63355]: INFO nova.scheduler.client.report [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Deleted allocations for instance abe6726e-7d6a-46e0-9273-de3e681311eb [ 644.139215] env[63355]: DEBUG nova.compute.manager [req-28285ae9-5ba3-4e20-8321-c6b438e07af8 req-c6094143-bb13-4138-98a7-9ff4b232c8e0 service nova] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Received event network-changed-6b8ca815-f009-4f5f-9095-66890ffdef10 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 644.139758] env[63355]: DEBUG nova.compute.manager [req-28285ae9-5ba3-4e20-8321-c6b438e07af8 req-c6094143-bb13-4138-98a7-9ff4b232c8e0 service nova] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Refreshing instance network info cache due to event network-changed-6b8ca815-f009-4f5f-9095-66890ffdef10. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 644.139978] env[63355]: DEBUG oslo_concurrency.lockutils [req-28285ae9-5ba3-4e20-8321-c6b438e07af8 req-c6094143-bb13-4138-98a7-9ff4b232c8e0 service nova] Acquiring lock "refresh_cache-bc1d7402-3651-488b-a431-2cb2ae987d32" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.140768] env[63355]: DEBUG oslo_concurrency.lockutils [req-28285ae9-5ba3-4e20-8321-c6b438e07af8 req-c6094143-bb13-4138-98a7-9ff4b232c8e0 service nova] Acquired lock "refresh_cache-bc1d7402-3651-488b-a431-2cb2ae987d32" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.140768] env[63355]: DEBUG nova.network.neutron [req-28285ae9-5ba3-4e20-8321-c6b438e07af8 req-c6094143-bb13-4138-98a7-9ff4b232c8e0 service nova] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Refreshing network info cache for port 6b8ca815-f009-4f5f-9095-66890ffdef10 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 644.313198] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.316039] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9aa5a131-255b-4a94-9e06-9582da09effd tempest-ServersAaction247Test-1947241823 tempest-ServersAaction247Test-1947241823-project-member] Lock "abe6726e-7d6a-46e0-9273-de3e681311eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.653s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.323451] env[63355]: ERROR nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6b8ca815-f009-4f5f-9095-66890ffdef10, please check neutron logs for more information. [ 644.323451] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 644.323451] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.323451] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 644.323451] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.323451] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 644.323451] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.323451] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 644.323451] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.323451] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 644.323451] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.323451] env[63355]: ERROR nova.compute.manager raise self.value [ 644.323451] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.323451] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 644.323451] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.323451] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 644.324023] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.324023] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 644.324023] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6b8ca815-f009-4f5f-9095-66890ffdef10, please check neutron logs for more information. [ 644.324023] env[63355]: ERROR nova.compute.manager [ 644.324023] env[63355]: Traceback (most recent call last): [ 644.324023] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 644.324023] env[63355]: listener.cb(fileno) [ 644.324023] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.324023] env[63355]: result = function(*args, **kwargs) [ 644.324023] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.324023] env[63355]: return func(*args, **kwargs) [ 644.324023] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.324023] env[63355]: raise e [ 644.324023] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.324023] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 644.324023] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.324023] env[63355]: created_port_ids = self._update_ports_for_instance( [ 644.324023] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.324023] env[63355]: with excutils.save_and_reraise_exception(): [ 644.324023] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.324023] env[63355]: self.force_reraise() [ 644.324023] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.324023] env[63355]: raise self.value [ 644.324023] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.324023] env[63355]: updated_port = self._update_port( [ 644.324023] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.324023] env[63355]: _ensure_no_port_binding_failure(port) [ 644.324023] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.324023] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 644.325056] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 6b8ca815-f009-4f5f-9095-66890ffdef10, please check neutron logs for more information. [ 644.325056] env[63355]: Removing descriptor: 18 [ 644.325056] env[63355]: ERROR nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6b8ca815-f009-4f5f-9095-66890ffdef10, please check neutron logs for more information. [ 644.325056] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Traceback (most recent call last): [ 644.325056] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 644.325056] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] yield resources [ 644.325056] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.325056] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] self.driver.spawn(context, instance, image_meta, [ 644.325056] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 644.325056] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.325056] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.325056] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] vm_ref = self.build_virtual_machine(instance, [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] for vif in network_info: [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] return self._sync_wrapper(fn, *args, **kwargs) [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] self.wait() [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] self[:] = self._gt.wait() [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] return self._exit_event.wait() [ 644.325420] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] result = hub.switch() [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] return self.greenlet.switch() [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] result = function(*args, **kwargs) [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] return func(*args, **kwargs) [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] raise e [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] nwinfo = self.network_api.allocate_for_instance( [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.325974] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] created_port_ids = self._update_ports_for_instance( [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] with excutils.save_and_reraise_exception(): [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] self.force_reraise() [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] raise self.value [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] updated_port = self._update_port( [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] _ensure_no_port_binding_failure(port) [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.326439] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] raise exception.PortBindingFailed(port_id=port['id']) [ 644.327055] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] nova.exception.PortBindingFailed: Binding failed for port 6b8ca815-f009-4f5f-9095-66890ffdef10, please check neutron logs for more information. [ 644.327055] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] [ 644.327055] env[63355]: INFO nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Terminating instance [ 644.328295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Acquiring lock "refresh_cache-bc1d7402-3651-488b-a431-2cb2ae987d32" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.395459] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.660193] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11fba5b5-23fa-4e1a-8fc2-daa613838b7d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.663273] env[63355]: DEBUG nova.network.neutron [req-28285ae9-5ba3-4e20-8321-c6b438e07af8 req-c6094143-bb13-4138-98a7-9ff4b232c8e0 service nova] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.669597] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb488423-874c-404d-a59c-f037395193d6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.699268] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356deb4d-2966-485b-b997-7b6af83e9252 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.706330] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df77c1e5-5c5b-4585-a4bf-aa9ee3256a02 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.710480] env[63355]: DEBUG nova.network.neutron [req-28285ae9-5ba3-4e20-8321-c6b438e07af8 req-c6094143-bb13-4138-98a7-9ff4b232c8e0 service nova] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.721046] env[63355]: DEBUG nova.compute.provider_tree [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.898223] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Releasing lock "refresh_cache-5c0ba7c6-9946-4765-be4b-d48d3823a68d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.899097] env[63355]: DEBUG nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 644.899097] env[63355]: DEBUG nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.899097] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 644.914365] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.213858] env[63355]: DEBUG oslo_concurrency.lockutils [req-28285ae9-5ba3-4e20-8321-c6b438e07af8 req-c6094143-bb13-4138-98a7-9ff4b232c8e0 service nova] Releasing lock "refresh_cache-bc1d7402-3651-488b-a431-2cb2ae987d32" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.214301] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Acquired lock "refresh_cache-bc1d7402-3651-488b-a431-2cb2ae987d32" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.214489] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.223940] env[63355]: DEBUG nova.scheduler.client.report [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.418027] env[63355]: DEBUG nova.network.neutron [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.730447] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.731103] env[63355]: ERROR nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a3f82096-5d04-47e8-9c62-dedd206b67ed, please check neutron logs for more information. [ 645.731103] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Traceback (most recent call last): [ 645.731103] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.731103] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] self.driver.spawn(context, instance, image_meta, [ 645.731103] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 645.731103] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.731103] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.731103] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] vm_ref = self.build_virtual_machine(instance, [ 645.731103] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.731103] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.731103] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] for vif in network_info: [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] return self._sync_wrapper(fn, *args, **kwargs) [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] self.wait() [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] self[:] = self._gt.wait() [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] return self._exit_event.wait() [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] result = hub.switch() [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.731462] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] return self.greenlet.switch() [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] result = function(*args, **kwargs) [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] return func(*args, **kwargs) [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] raise e [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] nwinfo = self.network_api.allocate_for_instance( [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] created_port_ids = self._update_ports_for_instance( [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] with excutils.save_and_reraise_exception(): [ 645.731840] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] self.force_reraise() [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] raise self.value [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] updated_port = self._update_port( [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] _ensure_no_port_binding_failure(port) [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] raise exception.PortBindingFailed(port_id=port['id']) [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] nova.exception.PortBindingFailed: Binding failed for port a3f82096-5d04-47e8-9c62-dedd206b67ed, please check neutron logs for more information. [ 645.732280] env[63355]: ERROR nova.compute.manager [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] [ 645.732823] env[63355]: DEBUG nova.compute.utils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Binding failed for port a3f82096-5d04-47e8-9c62-dedd206b67ed, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 645.734067] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.736063] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.009s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.737926] env[63355]: INFO nova.compute.claims [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.742106] env[63355]: DEBUG nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Build of instance 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce was re-scheduled: Binding failed for port a3f82096-5d04-47e8-9c62-dedd206b67ed, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 645.742573] env[63355]: DEBUG nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 645.742803] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Acquiring lock "refresh_cache-1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.742948] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Acquired lock "refresh_cache-1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.743123] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.810892] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.920509] env[63355]: INFO nova.compute.manager [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] [instance: 5c0ba7c6-9946-4765-be4b-d48d3823a68d] Took 1.02 seconds to deallocate network for instance. [ 646.184339] env[63355]: DEBUG nova.compute.manager [req-debb4f24-db91-4e14-9e31-ae626c5dc1a9 req-323dc7fb-e061-4ddb-9d52-e28772047f87 service nova] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Received event network-vif-deleted-6b8ca815-f009-4f5f-9095-66890ffdef10 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 646.274820] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.313825] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Releasing lock "refresh_cache-bc1d7402-3651-488b-a431-2cb2ae987d32" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.314231] env[63355]: DEBUG nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 646.314445] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 646.315087] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d2290b1-62c6-4b1b-9815-39416fb21e5b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.328642] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5e45dc-5fea-4e53-bc5e-75abd9abc743 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.354310] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bc1d7402-3651-488b-a431-2cb2ae987d32 could not be found. [ 646.354541] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 646.354720] env[63355]: INFO nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Took 0.04 seconds to destroy the instance on the hypervisor. [ 646.354961] env[63355]: DEBUG oslo.service.loopingcall [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 646.355229] env[63355]: DEBUG nova.compute.manager [-] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.355329] env[63355]: DEBUG nova.network.neutron [-] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.372913] env[63355]: DEBUG nova.network.neutron [-] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.410136] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.875066] env[63355]: DEBUG nova.network.neutron [-] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.912906] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Releasing lock "refresh_cache-1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.913187] env[63355]: DEBUG nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 646.913854] env[63355]: DEBUG nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.914130] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.941401] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.952134] env[63355]: INFO nova.scheduler.client.report [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Deleted allocations for instance 5c0ba7c6-9946-4765-be4b-d48d3823a68d [ 647.268608] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a7f2b1d-527c-4de6-8464-021b30b64816 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.276520] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2b119a-a9da-46a5-afa3-dc6a8e166d83 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.314091] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01ed89d-aff4-47c8-833a-3d9861c59b1c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.321566] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9af15b9-eb98-48ca-897f-9b8d0acdb69c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.334538] env[63355]: DEBUG nova.compute.provider_tree [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.379578] env[63355]: INFO nova.compute.manager [-] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Took 1.02 seconds to deallocate network for instance. [ 647.380390] env[63355]: DEBUG nova.compute.claims [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 647.380726] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.446628] env[63355]: DEBUG nova.network.neutron [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.460729] env[63355]: DEBUG oslo_concurrency.lockutils [None req-05fe12b8-1152-42ff-b5c6-5ccf4bccd982 tempest-ServersTestManualDisk-42518727 tempest-ServersTestManualDisk-42518727-project-member] Lock "5c0ba7c6-9946-4765-be4b-d48d3823a68d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.127s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.840060] env[63355]: DEBUG nova.scheduler.client.report [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.949514] env[63355]: INFO nova.compute.manager [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] [instance: 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce] Took 1.03 seconds to deallocate network for instance. [ 647.965547] env[63355]: DEBUG nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.345966] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.345966] env[63355]: DEBUG nova.compute.manager [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 648.351466] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.763s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.353116] env[63355]: INFO nova.compute.claims [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.488299] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.859389] env[63355]: DEBUG nova.compute.utils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.868966] env[63355]: DEBUG nova.compute.manager [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Not allocating networking since 'none' was specified. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 648.899786] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Acquiring lock "05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.901123] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Lock "05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.982134] env[63355]: INFO nova.scheduler.client.report [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Deleted allocations for instance 1c68f354-9e0e-4e7e-925e-d31fab4ac1ce [ 649.372041] env[63355]: DEBUG nova.compute.manager [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 649.502302] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Acquiring lock "282f1b55-8a74-467a-9ec8-0684eb780970" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.503802] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Lock "282f1b55-8a74-467a-9ec8-0684eb780970" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.503802] env[63355]: DEBUG oslo_concurrency.lockutils [None req-00047ca9-85d3-41b9-bd41-388887d30186 tempest-ServersWithSpecificFlavorTestJSON-1344412689 tempest-ServersWithSpecificFlavorTestJSON-1344412689-project-member] Lock "1c68f354-9e0e-4e7e-925e-d31fab4ac1ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.758s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.865192] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e73ba1-11c8-432d-9f4a-4e631424adc9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.873370] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4a8df7-20b2-423e-bc86-05047e4b2016 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.906221] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f566d844-0e96-479e-b377-9d81120519c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.913910] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f12a7b5-4a09-42bd-9062-11bc270530ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.928724] env[63355]: DEBUG nova.compute.provider_tree [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.008466] env[63355]: DEBUG nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 650.406954] env[63355]: DEBUG nova.compute.manager [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 650.433579] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 650.433890] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 650.434126] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 650.434337] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 650.434510] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 650.434711] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 650.434963] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 650.435180] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 650.435384] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 650.435580] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 650.435782] env[63355]: DEBUG nova.virt.hardware [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.436677] env[63355]: DEBUG nova.scheduler.client.report [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.440241] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b9bb75-2704-4fb5-8d18-924f9c188bcf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.448813] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4956e456-4e54-429f-b56c-bf4012625662 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.464386] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 650.469998] env[63355]: DEBUG oslo.service.loopingcall [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.470569] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 650.470822] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1389ce4-add1-4767-abe9-f4c4e02dd995 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.488281] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 650.488281] env[63355]: value = "task-1349500" [ 650.488281] env[63355]: _type = "Task" [ 650.488281] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.495883] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349500, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.533932] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.944937] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.945502] env[63355]: DEBUG nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 650.948904] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.739s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.999342] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349500, 'name': CreateVM_Task, 'duration_secs': 0.340533} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.999505] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 651.000759] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.000759] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.000759] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 651.000910] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6ec3c2f-800c-45d8-ad02-fcd5477d619f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.006065] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 651.006065] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528cbfa0-5f6a-f967-e7b0-c68d5e0cc630" [ 651.006065] env[63355]: _type = "Task" [ 651.006065] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.013920] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528cbfa0-5f6a-f967-e7b0-c68d5e0cc630, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.456810] env[63355]: DEBUG nova.compute.utils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 651.461049] env[63355]: DEBUG nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 651.461228] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 651.516141] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528cbfa0-5f6a-f967-e7b0-c68d5e0cc630, 'name': SearchDatastore_Task, 'duration_secs': 0.008737} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.518662] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.518889] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 651.519125] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.519266] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.519439] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 651.520052] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-723d5ff4-1961-4483-871f-98663172f9cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.527540] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 651.527712] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 651.528435] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8959ff85-2a5b-4125-85f9-6dba90c488f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.536834] env[63355]: DEBUG nova.policy [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b772ef430bd24520b5c2b50e558466ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '250f17ca46ff47e9a8f779a16bbc4662', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 651.540025] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 651.540025] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d30ea9-100d-e4cc-2a9d-6206f476d637" [ 651.540025] env[63355]: _type = "Task" [ 651.540025] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.547730] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d30ea9-100d-e4cc-2a9d-6206f476d637, 'name': SearchDatastore_Task} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.551893] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50aa5b63-611c-4dde-a85f-18ceffb53280 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.556803] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 651.556803] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5252b7d8-b4ed-6df9-e759-49fd10867744" [ 651.556803] env[63355]: _type = "Task" [ 651.556803] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.564692] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5252b7d8-b4ed-6df9-e759-49fd10867744, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.961418] env[63355]: DEBUG nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 651.975272] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Successfully created port: 859440b2-5360-450e-9894-afa75c885899 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 651.980842] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c5e5b4-4b31-486f-8f16-87b73ff005ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.990665] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb537b97-e51d-49b1-a4e6-842e41bb6390 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.029113] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe2bd30-c7fb-412e-88c1-8de5a545e309 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.036772] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb62676-52c9-4669-b476-d7b9a66d2fcd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.054133] env[63355]: DEBUG nova.compute.provider_tree [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.069276] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5252b7d8-b4ed-6df9-e759-49fd10867744, 'name': SearchDatastore_Task, 'duration_secs': 0.008301} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.070151] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.070434] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] f9c8ed86-506b-4654-8e7e-cb218605764b/f9c8ed86-506b-4654-8e7e-cb218605764b.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 652.070692] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d61153f-34b5-4ee1-9373-cdf6c2ed8ce9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.077578] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 652.077578] env[63355]: value = "task-1349501" [ 652.077578] env[63355]: _type = "Task" [ 652.077578] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.087227] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349501, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.557456] env[63355]: DEBUG nova.scheduler.client.report [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.587941] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349501, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465928} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.588332] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] f9c8ed86-506b-4654-8e7e-cb218605764b/f9c8ed86-506b-4654-8e7e-cb218605764b.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 652.588606] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 652.588905] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0607043f-d6b2-4dad-a442-1fa3fe057797 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.601038] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 652.601038] env[63355]: value = "task-1349502" [ 652.601038] env[63355]: _type = "Task" [ 652.601038] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.607990] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349502, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.979888] env[63355]: DEBUG nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 653.034797] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 653.035093] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 653.035243] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 653.035452] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 653.035610] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 653.035849] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 653.036032] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 653.036205] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 653.036371] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 653.036463] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 653.036658] env[63355]: DEBUG nova.virt.hardware [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.037537] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b758f0c9-c3e3-4298-b16f-7187defe88e5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.050016] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71384874-944b-45c1-ab43-72d0bca6a91c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.063737] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.114s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.064510] env[63355]: ERROR nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b9ffc5f0-96d2-432e-8217-446b9a457ec3, please check neutron logs for more information. [ 653.064510] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Traceback (most recent call last): [ 653.064510] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.064510] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] self.driver.spawn(context, instance, image_meta, [ 653.064510] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 653.064510] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.064510] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.064510] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] vm_ref = self.build_virtual_machine(instance, [ 653.064510] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.064510] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.064510] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] for vif in network_info: [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] return self._sync_wrapper(fn, *args, **kwargs) [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] self.wait() [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] self[:] = self._gt.wait() [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] return self._exit_event.wait() [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] result = hub.switch() [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.065089] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] return self.greenlet.switch() [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] result = function(*args, **kwargs) [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] return func(*args, **kwargs) [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] raise e [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] nwinfo = self.network_api.allocate_for_instance( [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] created_port_ids = self._update_ports_for_instance( [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] with excutils.save_and_reraise_exception(): [ 653.065616] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] self.force_reraise() [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] raise self.value [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] updated_port = self._update_port( [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] _ensure_no_port_binding_failure(port) [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] raise exception.PortBindingFailed(port_id=port['id']) [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] nova.exception.PortBindingFailed: Binding failed for port b9ffc5f0-96d2-432e-8217-446b9a457ec3, please check neutron logs for more information. [ 653.066084] env[63355]: ERROR nova.compute.manager [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] [ 653.066564] env[63355]: DEBUG nova.compute.utils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Binding failed for port b9ffc5f0-96d2-432e-8217-446b9a457ec3, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 653.066944] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.444s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.070396] env[63355]: DEBUG nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Build of instance c80d746b-ad37-43d7-874e-1c1c848546f1 was re-scheduled: Binding failed for port b9ffc5f0-96d2-432e-8217-446b9a457ec3, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 653.070864] env[63355]: DEBUG nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 653.071133] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Acquiring lock "refresh_cache-c80d746b-ad37-43d7-874e-1c1c848546f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.071290] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Acquired lock "refresh_cache-c80d746b-ad37-43d7-874e-1c1c848546f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.071449] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.113661] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349502, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057547} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.113661] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 653.115253] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced0d662-dadd-4ff6-aff6-b50890f74e50 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.151056] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] f9c8ed86-506b-4654-8e7e-cb218605764b/f9c8ed86-506b-4654-8e7e-cb218605764b.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 653.151056] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48c53305-081c-4adc-b885-fb50da76a49c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.180890] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 653.180890] env[63355]: value = "task-1349503" [ 653.180890] env[63355]: _type = "Task" [ 653.180890] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.192393] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349503, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.290775] env[63355]: ERROR nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 859440b2-5360-450e-9894-afa75c885899, please check neutron logs for more information. [ 653.290775] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 653.290775] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.290775] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 653.290775] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.290775] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 653.290775] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.290775] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 653.290775] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.290775] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 653.290775] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.290775] env[63355]: ERROR nova.compute.manager raise self.value [ 653.290775] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.290775] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 653.290775] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.290775] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 653.291531] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.291531] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 653.291531] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 859440b2-5360-450e-9894-afa75c885899, please check neutron logs for more information. [ 653.291531] env[63355]: ERROR nova.compute.manager [ 653.291531] env[63355]: Traceback (most recent call last): [ 653.291531] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 653.291531] env[63355]: listener.cb(fileno) [ 653.291531] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.291531] env[63355]: result = function(*args, **kwargs) [ 653.291531] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.291531] env[63355]: return func(*args, **kwargs) [ 653.291531] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.291531] env[63355]: raise e [ 653.291531] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.291531] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 653.291531] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.291531] env[63355]: created_port_ids = self._update_ports_for_instance( [ 653.291531] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.291531] env[63355]: with excutils.save_and_reraise_exception(): [ 653.291531] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.291531] env[63355]: self.force_reraise() [ 653.291531] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.291531] env[63355]: raise self.value [ 653.291531] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.291531] env[63355]: updated_port = self._update_port( [ 653.291531] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.291531] env[63355]: _ensure_no_port_binding_failure(port) [ 653.291531] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.291531] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 653.292771] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 859440b2-5360-450e-9894-afa75c885899, please check neutron logs for more information. [ 653.292771] env[63355]: Removing descriptor: 17 [ 653.292771] env[63355]: ERROR nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 859440b2-5360-450e-9894-afa75c885899, please check neutron logs for more information. [ 653.292771] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Traceback (most recent call last): [ 653.292771] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 653.292771] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] yield resources [ 653.292771] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.292771] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] self.driver.spawn(context, instance, image_meta, [ 653.292771] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 653.292771] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.292771] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.292771] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] vm_ref = self.build_virtual_machine(instance, [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] for vif in network_info: [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] return self._sync_wrapper(fn, *args, **kwargs) [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] self.wait() [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] self[:] = self._gt.wait() [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] return self._exit_event.wait() [ 653.293454] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] result = hub.switch() [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] return self.greenlet.switch() [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] result = function(*args, **kwargs) [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] return func(*args, **kwargs) [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] raise e [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] nwinfo = self.network_api.allocate_for_instance( [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.294123] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] created_port_ids = self._update_ports_for_instance( [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] with excutils.save_and_reraise_exception(): [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] self.force_reraise() [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] raise self.value [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] updated_port = self._update_port( [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] _ensure_no_port_binding_failure(port) [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.298053] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] raise exception.PortBindingFailed(port_id=port['id']) [ 653.298800] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] nova.exception.PortBindingFailed: Binding failed for port 859440b2-5360-450e-9894-afa75c885899, please check neutron logs for more information. [ 653.298800] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] [ 653.298800] env[63355]: INFO nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Terminating instance [ 653.298800] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Acquiring lock "refresh_cache-1de0e894-1edb-4296-9e9b-2c1eb0c7c275" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.298800] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Acquired lock "refresh_cache-1de0e894-1edb-4296-9e9b-2c1eb0c7c275" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.298800] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.439462] env[63355]: DEBUG nova.compute.manager [req-8079f82a-bcbd-4e13-96f7-14cfb9dd4c47 req-1a07c79d-941a-41f5-b2b8-65675bded0b0 service nova] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Received event network-changed-859440b2-5360-450e-9894-afa75c885899 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.439851] env[63355]: DEBUG nova.compute.manager [req-8079f82a-bcbd-4e13-96f7-14cfb9dd4c47 req-1a07c79d-941a-41f5-b2b8-65675bded0b0 service nova] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Refreshing instance network info cache due to event network-changed-859440b2-5360-450e-9894-afa75c885899. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 653.440124] env[63355]: DEBUG oslo_concurrency.lockutils [req-8079f82a-bcbd-4e13-96f7-14cfb9dd4c47 req-1a07c79d-941a-41f5-b2b8-65675bded0b0 service nova] Acquiring lock "refresh_cache-1de0e894-1edb-4296-9e9b-2c1eb0c7c275" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.595580] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.646753] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.690173] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349503, 'name': ReconfigVM_Task, 'duration_secs': 0.285318} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.692505] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Reconfigured VM instance instance-00000011 to attach disk [datastore2] f9c8ed86-506b-4654-8e7e-cb218605764b/f9c8ed86-506b-4654-8e7e-cb218605764b.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 653.693854] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d34858a7-b0a7-41c9-829b-d5fc7e67da88 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.699706] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 653.699706] env[63355]: value = "task-1349504" [ 653.699706] env[63355]: _type = "Task" [ 653.699706] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.708053] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349504, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.828480] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.986151] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.032168] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574d801c-a372-47d4-a62c-58560fe4a632 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.041359] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27367375-f427-44a1-a3e5-934ae150b98a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.072781] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4d5f16-e9ef-4df1-957c-004d5988ff66 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.081160] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd16135-b7ae-414b-9149-110de541619b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.094431] env[63355]: DEBUG nova.compute.provider_tree [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.151465] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Releasing lock "refresh_cache-c80d746b-ad37-43d7-874e-1c1c848546f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.151862] env[63355]: DEBUG nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 654.151940] env[63355]: DEBUG nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.152106] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 654.172549] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.212051] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349504, 'name': Rename_Task, 'duration_secs': 0.138932} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.212693] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 654.213533] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-032c5dbb-dac4-47ef-a94e-28e04c1189d6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.220426] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 654.220426] env[63355]: value = "task-1349505" [ 654.220426] env[63355]: _type = "Task" [ 654.220426] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.229328] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.492271] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Releasing lock "refresh_cache-1de0e894-1edb-4296-9e9b-2c1eb0c7c275" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.492724] env[63355]: DEBUG nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 654.492917] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 654.494380] env[63355]: DEBUG oslo_concurrency.lockutils [req-8079f82a-bcbd-4e13-96f7-14cfb9dd4c47 req-1a07c79d-941a-41f5-b2b8-65675bded0b0 service nova] Acquired lock "refresh_cache-1de0e894-1edb-4296-9e9b-2c1eb0c7c275" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.494380] env[63355]: DEBUG nova.network.neutron [req-8079f82a-bcbd-4e13-96f7-14cfb9dd4c47 req-1a07c79d-941a-41f5-b2b8-65675bded0b0 service nova] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Refreshing network info cache for port 859440b2-5360-450e-9894-afa75c885899 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 654.494659] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50c245c7-1a2c-4197-9a60-fb7053b1cf73 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.504769] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04f7d5c-d4ee-408e-9e3d-8a04a286f1ce {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.528464] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1de0e894-1edb-4296-9e9b-2c1eb0c7c275 could not be found. [ 654.528721] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 654.528964] env[63355]: INFO nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Took 0.04 seconds to destroy the instance on the hypervisor. [ 654.529221] env[63355]: DEBUG oslo.service.loopingcall [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 654.533021] env[63355]: DEBUG nova.compute.manager [-] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.533021] env[63355]: DEBUG nova.network.neutron [-] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 654.542930] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.543179] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.549755] env[63355]: DEBUG nova.network.neutron [-] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.575148] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "795b61f6-a0f3-4aab-94b3-2540ccbb7312" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.575378] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "795b61f6-a0f3-4aab-94b3-2540ccbb7312" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.597587] env[63355]: DEBUG nova.scheduler.client.report [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.704474] env[63355]: DEBUG nova.network.neutron [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.731996] env[63355]: DEBUG oslo_vmware.api [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349505, 'name': PowerOnVM_Task, 'duration_secs': 0.421627} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.732336] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 654.732567] env[63355]: INFO nova.compute.manager [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Took 4.33 seconds to spawn the instance on the hypervisor. [ 654.732723] env[63355]: DEBUG nova.compute.manager [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 654.733609] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d20ad88-5313-42cb-a2fa-01e4ec0f7d20 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.015977] env[63355]: DEBUG nova.network.neutron [req-8079f82a-bcbd-4e13-96f7-14cfb9dd4c47 req-1a07c79d-941a-41f5-b2b8-65675bded0b0 service nova] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.053671] env[63355]: DEBUG nova.network.neutron [-] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.108850] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.039s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.108850] env[63355]: ERROR nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 761cdfb5-da3a-45b9-bba7-b18ba3eac6db, please check neutron logs for more information. [ 655.108850] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Traceback (most recent call last): [ 655.108850] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.108850] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] self.driver.spawn(context, instance, image_meta, [ 655.108850] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 655.108850] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.108850] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.108850] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] vm_ref = self.build_virtual_machine(instance, [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] for vif in network_info: [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] return self._sync_wrapper(fn, *args, **kwargs) [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] self.wait() [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] self[:] = self._gt.wait() [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] return self._exit_event.wait() [ 655.109110] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] result = hub.switch() [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] return self.greenlet.switch() [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] result = function(*args, **kwargs) [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] return func(*args, **kwargs) [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] raise e [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] nwinfo = self.network_api.allocate_for_instance( [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.109383] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] created_port_ids = self._update_ports_for_instance( [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] with excutils.save_and_reraise_exception(): [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] self.force_reraise() [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] raise self.value [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] updated_port = self._update_port( [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] _ensure_no_port_binding_failure(port) [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.109638] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] raise exception.PortBindingFailed(port_id=port['id']) [ 655.109961] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] nova.exception.PortBindingFailed: Binding failed for port 761cdfb5-da3a-45b9-bba7-b18ba3eac6db, please check neutron logs for more information. [ 655.109961] env[63355]: ERROR nova.compute.manager [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] [ 655.109961] env[63355]: DEBUG nova.compute.utils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Binding failed for port 761cdfb5-da3a-45b9-bba7-b18ba3eac6db, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 655.111069] env[63355]: DEBUG nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Build of instance 591a7907-e4a7-41d6-9cd0-5cb359f08810 was re-scheduled: Binding failed for port 761cdfb5-da3a-45b9-bba7-b18ba3eac6db, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 655.111508] env[63355]: DEBUG nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 655.111739] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Acquiring lock "refresh_cache-591a7907-e4a7-41d6-9cd0-5cb359f08810" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.111884] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Acquired lock "refresh_cache-591a7907-e4a7-41d6-9cd0-5cb359f08810" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.112055] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.113774] env[63355]: DEBUG nova.network.neutron [req-8079f82a-bcbd-4e13-96f7-14cfb9dd4c47 req-1a07c79d-941a-41f5-b2b8-65675bded0b0 service nova] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.114747] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.888s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.116210] env[63355]: INFO nova.compute.claims [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.177718] env[63355]: INFO nova.compute.manager [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] [instance: c80d746b-ad37-43d7-874e-1c1c848546f1] Took 1.03 seconds to deallocate network for instance. [ 655.254671] env[63355]: INFO nova.compute.manager [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Took 37.56 seconds to build instance. [ 655.537764] env[63355]: DEBUG nova.compute.manager [req-5e52ce99-f62c-4e7e-80cc-f7fc89cddd46 req-9513363b-88dd-48dc-a5dc-c60c891d04c2 service nova] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Received event network-vif-deleted-859440b2-5360-450e-9894-afa75c885899 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.556653] env[63355]: INFO nova.compute.manager [-] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Took 1.03 seconds to deallocate network for instance. [ 655.559192] env[63355]: DEBUG nova.compute.claims [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 655.559378] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.619652] env[63355]: DEBUG oslo_concurrency.lockutils [req-8079f82a-bcbd-4e13-96f7-14cfb9dd4c47 req-1a07c79d-941a-41f5-b2b8-65675bded0b0 service nova] Releasing lock "refresh_cache-1de0e894-1edb-4296-9e9b-2c1eb0c7c275" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.640463] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.738827] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.760416] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1753c60f-684b-40c9-9056-b6f84a02ee54 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "f9c8ed86-506b-4654-8e7e-cb218605764b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.199s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.986984] env[63355]: INFO nova.compute.manager [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Rebuilding instance [ 656.042732] env[63355]: DEBUG nova.compute.manager [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 656.043705] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087c1911-2739-4153-a564-7b6c7006d884 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.206967] env[63355]: INFO nova.scheduler.client.report [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Deleted allocations for instance c80d746b-ad37-43d7-874e-1c1c848546f1 [ 656.241542] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Releasing lock "refresh_cache-591a7907-e4a7-41d6-9cd0-5cb359f08810" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.241844] env[63355]: DEBUG nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 656.242095] env[63355]: DEBUG nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.242311] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.265129] env[63355]: DEBUG nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 656.268016] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.544155] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ad7a33-9949-4735-896e-4020c9fd8a89 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.551994] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c6b51e-be5b-4e37-9c63-f6a750f0f1f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.557831] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 656.558140] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c98ebc2-b8a2-48ee-92d1-3e3ecc1d6fea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.588975] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2715fd-15c9-4ecd-a365-42052eaa5bd6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.591688] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 656.591688] env[63355]: value = "task-1349506" [ 656.591688] env[63355]: _type = "Task" [ 656.591688] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.598546] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153998c1-5a84-46d2-8655-0fd9b9c8994b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.605330] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349506, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.617095] env[63355]: DEBUG nova.compute.provider_tree [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.719264] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa36dd72-3a7a-42ec-8b94-d2b812c3121b tempest-ServerDiagnosticsTest-625634034 tempest-ServerDiagnosticsTest-625634034-project-member] Lock "c80d746b-ad37-43d7-874e-1c1c848546f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.596s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.770556] env[63355]: DEBUG nova.network.neutron [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.786639] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.101917] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349506, 'name': PowerOffVM_Task, 'duration_secs': 0.204602} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.101917] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 657.101917] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 657.102843] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2be93c0-bdc6-489d-88c7-6766b099c15e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.110047] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 657.111538] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0e3b6e8-0062-4f44-943b-55dc3f570bc4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.119979] env[63355]: DEBUG nova.scheduler.client.report [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.139104] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 657.139104] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 657.139104] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleting the datastore file [datastore2] f9c8ed86-506b-4654-8e7e-cb218605764b {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 657.139104] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2359f607-708f-4cd0-91b1-147388337b80 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.145820] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 657.145820] env[63355]: value = "task-1349508" [ 657.145820] env[63355]: _type = "Task" [ 657.145820] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.159039] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349508, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.223236] env[63355]: DEBUG nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 657.274748] env[63355]: INFO nova.compute.manager [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] [instance: 591a7907-e4a7-41d6-9cd0-5cb359f08810] Took 1.03 seconds to deallocate network for instance. [ 657.626969] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.628243] env[63355]: DEBUG nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 657.631196] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.008s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.655080] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349508, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128703} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.655339] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 657.655523] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 657.655702] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 657.753545] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.138507] env[63355]: DEBUG nova.compute.utils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 658.143797] env[63355]: DEBUG nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 658.143797] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 658.236215] env[63355]: DEBUG nova.policy [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4cc824dbae1f40918c3c57fb739fcfee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8749867b3cb44cadb7096608072d4904', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 658.316815] env[63355]: INFO nova.scheduler.client.report [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Deleted allocations for instance 591a7907-e4a7-41d6-9cd0-5cb359f08810 [ 658.643964] env[63355]: DEBUG nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 658.653381] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f208460-8ce1-48c3-bf2e-3d0a845422fb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.667927] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6895194b-31cf-4eec-aa9b-d5ae1890f5f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.708342] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4734e1-cc15-4532-8895-bbed2d4959ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.718182] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7de576-eca5-48a2-bc44-ccffc4ac0d69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.726958] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 658.726958] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 658.726958] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 658.726958] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 658.727118] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 658.727118] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 658.727118] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 658.727118] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 658.727118] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 658.727233] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 658.727233] env[63355]: DEBUG nova.virt.hardware [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 658.727233] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9ef0b2-783c-41ba-b0ba-1ede01cbd81e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.740589] env[63355]: DEBUG nova.compute.provider_tree [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.745263] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092a8e1a-2d71-4d84-bfa4-004fe524aa7c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.760496] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 658.766299] env[63355]: DEBUG oslo.service.loopingcall [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.766776] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 658.766984] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ad933e42-59a8-4c14-ab03-a6644a6d75cc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.785063] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 658.785063] env[63355]: value = "task-1349509" [ 658.785063] env[63355]: _type = "Task" [ 658.785063] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.792521] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349509, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.826780] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ad7c2d73-7688-413e-9e44-449773e4c121 tempest-ServerMetadataTestJSON-1267844983 tempest-ServerMetadataTestJSON-1267844983-project-member] Lock "591a7907-e4a7-41d6-9cd0-5cb359f08810" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.605s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.902289] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Successfully created port: 383c4b03-ec5b-4ef4-8b72-0dc09535c971 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 659.247199] env[63355]: DEBUG nova.scheduler.client.report [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.298652] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349509, 'name': CreateVM_Task, 'duration_secs': 0.248558} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.298822] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 659.299438] env[63355]: DEBUG oslo_concurrency.lockutils [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.299753] env[63355]: DEBUG oslo_concurrency.lockutils [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.300279] env[63355]: DEBUG oslo_concurrency.lockutils [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 659.300525] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55569683-fcbf-4f2c-b062-49fca12dfd7a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.306169] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 659.306169] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5284285e-e1f8-d2c6-4e97-fcc5950b0b08" [ 659.306169] env[63355]: _type = "Task" [ 659.306169] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.315497] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5284285e-e1f8-d2c6-4e97-fcc5950b0b08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.330937] env[63355]: DEBUG nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 659.660932] env[63355]: DEBUG nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 659.696430] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 659.696674] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 659.696822] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 659.696994] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 659.697159] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 659.697307] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 659.697585] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 659.697764] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 659.697925] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 659.698351] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 659.698453] env[63355]: DEBUG nova.virt.hardware [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 659.699348] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21da37a7-5b44-4b1e-b6b7-47ef2fae0e26 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.708038] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1133e9ee-3329-4b8c-87b3-78c9d04bad62 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.752031] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.120s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.752399] env[63355]: ERROR nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b60350f9-d151-4fea-8aad-7d681d5936b9, please check neutron logs for more information. [ 659.752399] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Traceback (most recent call last): [ 659.752399] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.752399] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] self.driver.spawn(context, instance, image_meta, [ 659.752399] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 659.752399] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.752399] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.752399] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] vm_ref = self.build_virtual_machine(instance, [ 659.752399] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.752399] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.752399] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] for vif in network_info: [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] return self._sync_wrapper(fn, *args, **kwargs) [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] self.wait() [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] self[:] = self._gt.wait() [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] return self._exit_event.wait() [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] result = hub.switch() [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 659.753604] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] return self.greenlet.switch() [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] result = function(*args, **kwargs) [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] return func(*args, **kwargs) [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] raise e [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] nwinfo = self.network_api.allocate_for_instance( [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] created_port_ids = self._update_ports_for_instance( [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] with excutils.save_and_reraise_exception(): [ 659.753956] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] self.force_reraise() [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] raise self.value [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] updated_port = self._update_port( [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] _ensure_no_port_binding_failure(port) [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] raise exception.PortBindingFailed(port_id=port['id']) [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] nova.exception.PortBindingFailed: Binding failed for port b60350f9-d151-4fea-8aad-7d681d5936b9, please check neutron logs for more information. [ 659.754251] env[63355]: ERROR nova.compute.manager [instance: 4392313f-2fed-4308-b79a-324b3431d5af] [ 659.754542] env[63355]: DEBUG nova.compute.utils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Binding failed for port b60350f9-d151-4fea-8aad-7d681d5936b9, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 659.754542] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.923s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.755860] env[63355]: INFO nova.compute.claims [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.758502] env[63355]: DEBUG nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Build of instance 4392313f-2fed-4308-b79a-324b3431d5af was re-scheduled: Binding failed for port b60350f9-d151-4fea-8aad-7d681d5936b9, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 659.758928] env[63355]: DEBUG nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 659.759814] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Acquiring lock "refresh_cache-4392313f-2fed-4308-b79a-324b3431d5af" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.759814] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Acquired lock "refresh_cache-4392313f-2fed-4308-b79a-324b3431d5af" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.759814] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.817944] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5284285e-e1f8-d2c6-4e97-fcc5950b0b08, 'name': SearchDatastore_Task, 'duration_secs': 0.011045} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.818415] env[63355]: DEBUG oslo_concurrency.lockutils [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.818856] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 659.819200] env[63355]: DEBUG oslo_concurrency.lockutils [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.819448] env[63355]: DEBUG oslo_concurrency.lockutils [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.819747] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 659.820317] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-669aed51-5d9f-4a7e-bdb1-7fcd5446d2bd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.828389] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 659.828666] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 659.831022] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d5b988a-07e9-485c-bc99-ffa833b329f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.835136] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 659.835136] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d79d47-35ec-fd76-6b51-98c5d8de81eb" [ 659.835136] env[63355]: _type = "Task" [ 659.835136] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.852039] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d79d47-35ec-fd76-6b51-98c5d8de81eb, 'name': SearchDatastore_Task, 'duration_secs': 0.007964} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.852039] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f25921e-c13e-4963-a186-fe64625f8b5f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.856026] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 659.856026] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52aceb90-b01e-5426-472e-383f288f63c9" [ 659.856026] env[63355]: _type = "Task" [ 659.856026] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.861328] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.864839] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52aceb90-b01e-5426-472e-383f288f63c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.021631] env[63355]: DEBUG nova.compute.manager [req-3a0d4257-0cb8-4615-b276-ab13b1347152 req-1b6898bf-c05e-4309-9191-41aba481cf09 service nova] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Received event network-changed-383c4b03-ec5b-4ef4-8b72-0dc09535c971 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.021821] env[63355]: DEBUG nova.compute.manager [req-3a0d4257-0cb8-4615-b276-ab13b1347152 req-1b6898bf-c05e-4309-9191-41aba481cf09 service nova] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Refreshing instance network info cache due to event network-changed-383c4b03-ec5b-4ef4-8b72-0dc09535c971. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 660.022052] env[63355]: DEBUG oslo_concurrency.lockutils [req-3a0d4257-0cb8-4615-b276-ab13b1347152 req-1b6898bf-c05e-4309-9191-41aba481cf09 service nova] Acquiring lock "refresh_cache-10112b62-f41e-4413-9bc8-c1306c5c1bc1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.022197] env[63355]: DEBUG oslo_concurrency.lockutils [req-3a0d4257-0cb8-4615-b276-ab13b1347152 req-1b6898bf-c05e-4309-9191-41aba481cf09 service nova] Acquired lock "refresh_cache-10112b62-f41e-4413-9bc8-c1306c5c1bc1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.022374] env[63355]: DEBUG nova.network.neutron [req-3a0d4257-0cb8-4615-b276-ab13b1347152 req-1b6898bf-c05e-4309-9191-41aba481cf09 service nova] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Refreshing network info cache for port 383c4b03-ec5b-4ef4-8b72-0dc09535c971 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 660.257458] env[63355]: ERROR nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 383c4b03-ec5b-4ef4-8b72-0dc09535c971, please check neutron logs for more information. [ 660.257458] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.257458] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.257458] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.257458] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.257458] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.257458] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.257458] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.257458] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.257458] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 660.257458] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.257458] env[63355]: ERROR nova.compute.manager raise self.value [ 660.257458] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.257458] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.257458] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.257458] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.258024] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.258024] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.258024] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 383c4b03-ec5b-4ef4-8b72-0dc09535c971, please check neutron logs for more information. [ 660.258024] env[63355]: ERROR nova.compute.manager [ 660.258024] env[63355]: Traceback (most recent call last): [ 660.258024] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.258024] env[63355]: listener.cb(fileno) [ 660.258024] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.258024] env[63355]: result = function(*args, **kwargs) [ 660.258024] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.258024] env[63355]: return func(*args, **kwargs) [ 660.258024] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.258024] env[63355]: raise e [ 660.258024] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.258024] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 660.258024] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.258024] env[63355]: created_port_ids = self._update_ports_for_instance( [ 660.258024] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.258024] env[63355]: with excutils.save_and_reraise_exception(): [ 660.258024] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.258024] env[63355]: self.force_reraise() [ 660.258024] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.258024] env[63355]: raise self.value [ 660.258024] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.258024] env[63355]: updated_port = self._update_port( [ 660.258024] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.258024] env[63355]: _ensure_no_port_binding_failure(port) [ 660.258024] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.258024] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.258617] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 383c4b03-ec5b-4ef4-8b72-0dc09535c971, please check neutron logs for more information. [ 660.258617] env[63355]: Removing descriptor: 17 [ 660.258617] env[63355]: ERROR nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 383c4b03-ec5b-4ef4-8b72-0dc09535c971, please check neutron logs for more information. [ 660.258617] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Traceback (most recent call last): [ 660.258617] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 660.258617] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] yield resources [ 660.258617] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.258617] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] self.driver.spawn(context, instance, image_meta, [ 660.258617] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 660.258617] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.258617] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.258617] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] vm_ref = self.build_virtual_machine(instance, [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] for vif in network_info: [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] return self._sync_wrapper(fn, *args, **kwargs) [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] self.wait() [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] self[:] = self._gt.wait() [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] return self._exit_event.wait() [ 660.258907] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] result = hub.switch() [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] return self.greenlet.switch() [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] result = function(*args, **kwargs) [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] return func(*args, **kwargs) [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] raise e [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] nwinfo = self.network_api.allocate_for_instance( [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.259212] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] created_port_ids = self._update_ports_for_instance( [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] with excutils.save_and_reraise_exception(): [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] self.force_reraise() [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] raise self.value [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] updated_port = self._update_port( [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] _ensure_no_port_binding_failure(port) [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.259635] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] raise exception.PortBindingFailed(port_id=port['id']) [ 660.259947] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] nova.exception.PortBindingFailed: Binding failed for port 383c4b03-ec5b-4ef4-8b72-0dc09535c971, please check neutron logs for more information. [ 660.259947] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] [ 660.259947] env[63355]: INFO nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Terminating instance [ 660.266337] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Acquiring lock "refresh_cache-10112b62-f41e-4413-9bc8-c1306c5c1bc1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.281627] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.369186] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52aceb90-b01e-5426-472e-383f288f63c9, 'name': SearchDatastore_Task, 'duration_secs': 0.008562} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.369186] env[63355]: DEBUG oslo_concurrency.lockutils [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.369186] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] f9c8ed86-506b-4654-8e7e-cb218605764b/f9c8ed86-506b-4654-8e7e-cb218605764b.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 660.369186] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3cda5ee2-fa7b-4e10-ab21-04ee514c39cb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.375114] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.377529] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 660.377529] env[63355]: value = "task-1349510" [ 660.377529] env[63355]: _type = "Task" [ 660.377529] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.389147] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349510, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.553406] env[63355]: DEBUG nova.network.neutron [req-3a0d4257-0cb8-4615-b276-ab13b1347152 req-1b6898bf-c05e-4309-9191-41aba481cf09 service nova] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.663743] env[63355]: DEBUG nova.network.neutron [req-3a0d4257-0cb8-4615-b276-ab13b1347152 req-1b6898bf-c05e-4309-9191-41aba481cf09 service nova] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.886313] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Releasing lock "refresh_cache-4392313f-2fed-4308-b79a-324b3431d5af" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.886672] env[63355]: DEBUG nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 660.886745] env[63355]: DEBUG nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.886875] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.894654] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349510, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484922} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.894654] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] f9c8ed86-506b-4654-8e7e-cb218605764b/f9c8ed86-506b-4654-8e7e-cb218605764b.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 660.894654] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 660.894654] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a017a95-91e5-4145-a7ae-9283d60f36df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.901465] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 660.901465] env[63355]: value = "task-1349511" [ 660.901465] env[63355]: _type = "Task" [ 660.901465] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.909077] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349511, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.922759] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.169770] env[63355]: DEBUG oslo_concurrency.lockutils [req-3a0d4257-0cb8-4615-b276-ab13b1347152 req-1b6898bf-c05e-4309-9191-41aba481cf09 service nova] Releasing lock "refresh_cache-10112b62-f41e-4413-9bc8-c1306c5c1bc1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.170959] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Acquired lock "refresh_cache-10112b62-f41e-4413-9bc8-c1306c5c1bc1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.170959] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 661.329890] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d67d008-9220-4770-a262-323a0dfc5d9e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.337139] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513afacb-3620-4a0b-8607-2b8b217782a0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.373693] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b76cff-213b-4065-8a5c-caade8c4686a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.381454] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e47c11-43e1-4907-9a94-5ef0d70e916d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.398412] env[63355]: DEBUG nova.compute.provider_tree [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.410381] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349511, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069614} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.410381] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 661.410653] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a02c26-af20-4f40-99e0-ec2060704a5c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.431497] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] f9c8ed86-506b-4654-8e7e-cb218605764b/f9c8ed86-506b-4654-8e7e-cb218605764b.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 661.431923] env[63355]: DEBUG nova.network.neutron [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.433379] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-769e5771-23ea-4765-946e-a0a86ce7eb9b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.455139] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 661.455139] env[63355]: value = "task-1349512" [ 661.455139] env[63355]: _type = "Task" [ 661.455139] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.462635] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349512, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.702543] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.820144] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.906781] env[63355]: DEBUG nova.scheduler.client.report [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.950255] env[63355]: INFO nova.compute.manager [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] [instance: 4392313f-2fed-4308-b79a-324b3431d5af] Took 1.06 seconds to deallocate network for instance. [ 661.968252] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349512, 'name': ReconfigVM_Task, 'duration_secs': 0.268268} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.968392] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Reconfigured VM instance instance-00000011 to attach disk [datastore2] f9c8ed86-506b-4654-8e7e-cb218605764b/f9c8ed86-506b-4654-8e7e-cb218605764b.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 661.968936] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d01dca7c-278f-4f92-8a02-78a6b917574c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.975995] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 661.975995] env[63355]: value = "task-1349513" [ 661.975995] env[63355]: _type = "Task" [ 661.975995] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.984763] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349513, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.179713] env[63355]: DEBUG nova.compute.manager [req-e744451f-1cfa-420b-9fbd-e899514f5e0e req-1be71967-4561-49ce-a936-179973f985c0 service nova] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Received event network-vif-deleted-383c4b03-ec5b-4ef4-8b72-0dc09535c971 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.324275] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Releasing lock "refresh_cache-10112b62-f41e-4413-9bc8-c1306c5c1bc1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.324785] env[63355]: DEBUG nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 662.324979] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 662.325296] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5ae4bff-ade9-45f8-b5c8-e427bfac5201 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.337584] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40eb8fdc-5bd2-47b3-b756-1387f95cc7c9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.365555] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 10112b62-f41e-4413-9bc8-c1306c5c1bc1 could not be found. [ 662.365794] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 662.365977] env[63355]: INFO nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 662.366434] env[63355]: DEBUG oslo.service.loopingcall [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.366434] env[63355]: DEBUG nova.compute.manager [-] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.366522] env[63355]: DEBUG nova.network.neutron [-] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 662.396563] env[63355]: DEBUG nova.network.neutron [-] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.419031] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.663s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.419031] env[63355]: DEBUG nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 662.425121] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.044s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.490296] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349513, 'name': Rename_Task, 'duration_secs': 0.1366} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.491106] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 662.491106] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ddf3246-80e0-4fa0-b358-8f7123e6b7e2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.498456] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 662.498456] env[63355]: value = "task-1349514" [ 662.498456] env[63355]: _type = "Task" [ 662.498456] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.513513] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349514, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.906203] env[63355]: DEBUG nova.network.neutron [-] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.929781] env[63355]: DEBUG nova.compute.utils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.933397] env[63355]: DEBUG nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.934718] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 663.006471] env[63355]: INFO nova.scheduler.client.report [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Deleted allocations for instance 4392313f-2fed-4308-b79a-324b3431d5af [ 663.028954] env[63355]: DEBUG oslo_vmware.api [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349514, 'name': PowerOnVM_Task, 'duration_secs': 0.43716} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.029741] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 663.029741] env[63355]: DEBUG nova.compute.manager [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 663.030811] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f783d70-513d-4143-bbd9-e0dce98239d2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.034922] env[63355]: DEBUG nova.policy [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ec72d25e81f4e1893d8865cc3f1b79c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1650edaa46e498fa233ec3b4c81b900', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.408612] env[63355]: INFO nova.compute.manager [-] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Took 1.04 seconds to deallocate network for instance. [ 663.413262] env[63355]: DEBUG nova.compute.claims [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 663.413435] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.438390] env[63355]: DEBUG nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 663.454775] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff811ccb-28bf-40b5-84d9-d0f58e9d0528 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.464473] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10bdeca-5bd6-4b36-9bca-e0477364872d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.505023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b78d239-02dc-4825-ba91-83a9c3a75eab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.512393] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5678e14-06c3-403e-9b9e-ee2c9873d6d0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.526806] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e8300e2-15ad-4ec8-a295-c508ac8ad9c9 tempest-TenantUsagesTestJSON-742280717 tempest-TenantUsagesTestJSON-742280717-project-member] Lock "4392313f-2fed-4308-b79a-324b3431d5af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.278s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.527359] env[63355]: DEBUG nova.compute.provider_tree [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.531976] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Successfully created port: c43c4079-a3a9-401e-ba17-aab8f7a656ad {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 663.559254] env[63355]: DEBUG oslo_concurrency.lockutils [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.033026] env[63355]: DEBUG nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 664.034301] env[63355]: DEBUG nova.scheduler.client.report [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.186666] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Acquiring lock "7816d808-c9dd-403f-b993-c6d9dc458c7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.186946] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Lock "7816d808-c9dd-403f-b993-c6d9dc458c7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.456344] env[63355]: DEBUG nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 664.487940] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 664.488245] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 664.488406] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 664.488591] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 664.488733] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 664.488878] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 664.489095] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 664.489255] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 664.489419] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 664.489579] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 664.489751] env[63355]: DEBUG nova.virt.hardware [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.490651] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e28aa1-27e7-4346-9bc8-c030ba492315 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.498782] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553f7933-4f49-45d6-86d5-8dd9dfbf3ef9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.543768] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.119s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.546110] env[63355]: ERROR nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6b8ca815-f009-4f5f-9095-66890ffdef10, please check neutron logs for more information. [ 664.546110] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Traceback (most recent call last): [ 664.546110] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.546110] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] self.driver.spawn(context, instance, image_meta, [ 664.546110] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 664.546110] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.546110] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.546110] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] vm_ref = self.build_virtual_machine(instance, [ 664.546110] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.546110] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.546110] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] for vif in network_info: [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] return self._sync_wrapper(fn, *args, **kwargs) [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] self.wait() [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] self[:] = self._gt.wait() [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] return self._exit_event.wait() [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] result = hub.switch() [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 664.547065] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] return self.greenlet.switch() [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] result = function(*args, **kwargs) [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] return func(*args, **kwargs) [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] raise e [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] nwinfo = self.network_api.allocate_for_instance( [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] created_port_ids = self._update_ports_for_instance( [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] with excutils.save_and_reraise_exception(): [ 664.547410] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] self.force_reraise() [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] raise self.value [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] updated_port = self._update_port( [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] _ensure_no_port_binding_failure(port) [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] raise exception.PortBindingFailed(port_id=port['id']) [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] nova.exception.PortBindingFailed: Binding failed for port 6b8ca815-f009-4f5f-9095-66890ffdef10, please check neutron logs for more information. [ 664.547691] env[63355]: ERROR nova.compute.manager [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] [ 664.547993] env[63355]: DEBUG nova.compute.utils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Binding failed for port 6b8ca815-f009-4f5f-9095-66890ffdef10, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 664.552213] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.064s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.553829] env[63355]: INFO nova.compute.claims [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.570614] env[63355]: DEBUG nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Build of instance bc1d7402-3651-488b-a431-2cb2ae987d32 was re-scheduled: Binding failed for port 6b8ca815-f009-4f5f-9095-66890ffdef10, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 664.570614] env[63355]: DEBUG nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 664.570614] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Acquiring lock "refresh_cache-bc1d7402-3651-488b-a431-2cb2ae987d32" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.570614] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Acquired lock "refresh_cache-bc1d7402-3651-488b-a431-2cb2ae987d32" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.570854] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.580643] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.875085] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "f9c8ed86-506b-4654-8e7e-cb218605764b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.875325] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "f9c8ed86-506b-4654-8e7e-cb218605764b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.875798] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "f9c8ed86-506b-4654-8e7e-cb218605764b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.876025] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "f9c8ed86-506b-4654-8e7e-cb218605764b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.876220] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "f9c8ed86-506b-4654-8e7e-cb218605764b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.881243] env[63355]: INFO nova.compute.manager [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Terminating instance [ 664.887474] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "refresh_cache-f9c8ed86-506b-4654-8e7e-cb218605764b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.887474] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired lock "refresh_cache-f9c8ed86-506b-4654-8e7e-cb218605764b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.887474] env[63355]: DEBUG nova.network.neutron [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.098046] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.183912] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.411175] env[63355]: DEBUG nova.network.neutron [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.518525] env[63355]: DEBUG nova.network.neutron [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.687419] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Releasing lock "refresh_cache-bc1d7402-3651-488b-a431-2cb2ae987d32" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.687665] env[63355]: DEBUG nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 665.687828] env[63355]: DEBUG nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.687992] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.709290] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.886756] env[63355]: DEBUG nova.compute.manager [req-dc6966db-fa51-40c7-8104-a3dfa354d1ae req-930c6d82-0713-495a-9cbe-8d4085dfb8c0 service nova] [instance: b275344f-695e-463b-8bfe-2bee97296349] Received event network-changed-c43c4079-a3a9-401e-ba17-aab8f7a656ad {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.889650] env[63355]: DEBUG nova.compute.manager [req-dc6966db-fa51-40c7-8104-a3dfa354d1ae req-930c6d82-0713-495a-9cbe-8d4085dfb8c0 service nova] [instance: b275344f-695e-463b-8bfe-2bee97296349] Refreshing instance network info cache due to event network-changed-c43c4079-a3a9-401e-ba17-aab8f7a656ad. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 665.889650] env[63355]: DEBUG oslo_concurrency.lockutils [req-dc6966db-fa51-40c7-8104-a3dfa354d1ae req-930c6d82-0713-495a-9cbe-8d4085dfb8c0 service nova] Acquiring lock "refresh_cache-b275344f-695e-463b-8bfe-2bee97296349" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.889650] env[63355]: DEBUG oslo_concurrency.lockutils [req-dc6966db-fa51-40c7-8104-a3dfa354d1ae req-930c6d82-0713-495a-9cbe-8d4085dfb8c0 service nova] Acquired lock "refresh_cache-b275344f-695e-463b-8bfe-2bee97296349" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.889650] env[63355]: DEBUG nova.network.neutron [req-dc6966db-fa51-40c7-8104-a3dfa354d1ae req-930c6d82-0713-495a-9cbe-8d4085dfb8c0 service nova] [instance: b275344f-695e-463b-8bfe-2bee97296349] Refreshing network info cache for port c43c4079-a3a9-401e-ba17-aab8f7a656ad {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 666.023574] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Releasing lock "refresh_cache-f9c8ed86-506b-4654-8e7e-cb218605764b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.023915] env[63355]: DEBUG nova.compute.manager [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 666.024121] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 666.029308] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2963d0b-3952-4bcf-9193-0018b23dcf46 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.039166] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 666.039166] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b842be94-fd2a-4f92-a9d3-61ff410ecb2f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.050609] env[63355]: DEBUG oslo_vmware.api [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 666.050609] env[63355]: value = "task-1349515" [ 666.050609] env[63355]: _type = "Task" [ 666.050609] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.068301] env[63355]: DEBUG oslo_vmware.api [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349515, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.174510] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ec2458-6023-4666-9649-8fad9330b0b6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.185716] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0703648a-9c05-4524-9226-d7aad4bc72e9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.222306] env[63355]: DEBUG nova.network.neutron [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.224594] env[63355]: ERROR nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c43c4079-a3a9-401e-ba17-aab8f7a656ad, please check neutron logs for more information. [ 666.224594] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 666.224594] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.224594] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 666.224594] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.224594] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 666.224594] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.224594] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 666.224594] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.224594] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 666.224594] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.224594] env[63355]: ERROR nova.compute.manager raise self.value [ 666.224594] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.224594] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 666.224594] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.224594] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 666.224982] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.224982] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 666.224982] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c43c4079-a3a9-401e-ba17-aab8f7a656ad, please check neutron logs for more information. [ 666.224982] env[63355]: ERROR nova.compute.manager [ 666.224982] env[63355]: Traceback (most recent call last): [ 666.224982] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 666.224982] env[63355]: listener.cb(fileno) [ 666.224982] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.224982] env[63355]: result = function(*args, **kwargs) [ 666.224982] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.224982] env[63355]: return func(*args, **kwargs) [ 666.224982] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.224982] env[63355]: raise e [ 666.224982] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.224982] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 666.224982] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.224982] env[63355]: created_port_ids = self._update_ports_for_instance( [ 666.224982] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.224982] env[63355]: with excutils.save_and_reraise_exception(): [ 666.224982] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.224982] env[63355]: self.force_reraise() [ 666.224982] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.224982] env[63355]: raise self.value [ 666.224982] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.224982] env[63355]: updated_port = self._update_port( [ 666.224982] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.224982] env[63355]: _ensure_no_port_binding_failure(port) [ 666.224982] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.224982] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 666.225767] env[63355]: nova.exception.PortBindingFailed: Binding failed for port c43c4079-a3a9-401e-ba17-aab8f7a656ad, please check neutron logs for more information. [ 666.225767] env[63355]: Removing descriptor: 17 [ 666.225767] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d457daaf-65d4-404d-b00f-27977272d7e5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.229087] env[63355]: ERROR nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c43c4079-a3a9-401e-ba17-aab8f7a656ad, please check neutron logs for more information. [ 666.229087] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] Traceback (most recent call last): [ 666.229087] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 666.229087] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] yield resources [ 666.229087] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.229087] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] self.driver.spawn(context, instance, image_meta, [ 666.229087] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 666.229087] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.229087] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.229087] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] vm_ref = self.build_virtual_machine(instance, [ 666.229087] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] for vif in network_info: [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] return self._sync_wrapper(fn, *args, **kwargs) [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] self.wait() [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] self[:] = self._gt.wait() [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] return self._exit_event.wait() [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.232172] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] result = hub.switch() [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] return self.greenlet.switch() [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] result = function(*args, **kwargs) [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] return func(*args, **kwargs) [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] raise e [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] nwinfo = self.network_api.allocate_for_instance( [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] created_port_ids = self._update_ports_for_instance( [ 666.232917] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] with excutils.save_and_reraise_exception(): [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] self.force_reraise() [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] raise self.value [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] updated_port = self._update_port( [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] _ensure_no_port_binding_failure(port) [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] raise exception.PortBindingFailed(port_id=port['id']) [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] nova.exception.PortBindingFailed: Binding failed for port c43c4079-a3a9-401e-ba17-aab8f7a656ad, please check neutron logs for more information. [ 666.233967] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] [ 666.234429] env[63355]: INFO nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Terminating instance [ 666.234429] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Acquiring lock "refresh_cache-b275344f-695e-463b-8bfe-2bee97296349" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.237892] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c5a608-b86c-4bda-b122-43711d639c5a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.254198] env[63355]: DEBUG nova.compute.provider_tree [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.417311] env[63355]: DEBUG nova.network.neutron [req-dc6966db-fa51-40c7-8104-a3dfa354d1ae req-930c6d82-0713-495a-9cbe-8d4085dfb8c0 service nova] [instance: b275344f-695e-463b-8bfe-2bee97296349] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.560646] env[63355]: DEBUG oslo_vmware.api [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349515, 'name': PowerOffVM_Task, 'duration_secs': 0.208617} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.560909] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 666.561180] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 666.561321] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fa8e8e4f-13ca-4bd5-89f6-0dfd1bf0c952 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.589819] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 666.589914] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 666.590053] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleting the datastore file [datastore2] f9c8ed86-506b-4654-8e7e-cb218605764b {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 666.590322] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7613a979-275a-4e71-b7d4-05f6a2ae65ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.597344] env[63355]: DEBUG oslo_vmware.api [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 666.597344] env[63355]: value = "task-1349517" [ 666.597344] env[63355]: _type = "Task" [ 666.597344] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.609164] env[63355]: DEBUG oslo_vmware.api [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349517, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.612842] env[63355]: DEBUG nova.network.neutron [req-dc6966db-fa51-40c7-8104-a3dfa354d1ae req-930c6d82-0713-495a-9cbe-8d4085dfb8c0 service nova] [instance: b275344f-695e-463b-8bfe-2bee97296349] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.732541] env[63355]: INFO nova.compute.manager [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] [instance: bc1d7402-3651-488b-a431-2cb2ae987d32] Took 1.04 seconds to deallocate network for instance. [ 666.758050] env[63355]: DEBUG nova.scheduler.client.report [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.109445] env[63355]: DEBUG oslo_vmware.api [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349517, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.110845} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.109787] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 667.110093] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 667.110158] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 667.110331] env[63355]: INFO nova.compute.manager [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Took 1.09 seconds to destroy the instance on the hypervisor. [ 667.110645] env[63355]: DEBUG oslo.service.loopingcall [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.110859] env[63355]: DEBUG nova.compute.manager [-] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.110960] env[63355]: DEBUG nova.network.neutron [-] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 667.116090] env[63355]: DEBUG oslo_concurrency.lockutils [req-dc6966db-fa51-40c7-8104-a3dfa354d1ae req-930c6d82-0713-495a-9cbe-8d4085dfb8c0 service nova] Releasing lock "refresh_cache-b275344f-695e-463b-8bfe-2bee97296349" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.116771] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Acquired lock "refresh_cache-b275344f-695e-463b-8bfe-2bee97296349" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.116956] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 667.130605] env[63355]: DEBUG nova.network.neutron [-] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.265563] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.713s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.268344] env[63355]: DEBUG nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 667.274488] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.740s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.275631] env[63355]: INFO nova.compute.claims [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.639570] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.639570] env[63355]: DEBUG nova.network.neutron [-] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.730159] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.771587] env[63355]: INFO nova.scheduler.client.report [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Deleted allocations for instance bc1d7402-3651-488b-a431-2cb2ae987d32 [ 667.782060] env[63355]: DEBUG nova.compute.utils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.788345] env[63355]: DEBUG nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 667.788345] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 667.839349] env[63355]: DEBUG nova.policy [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '843ccd4ab204481c956568280cf597f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a797f6acc3584178ae6c4c984c39af18', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 668.142126] env[63355]: INFO nova.compute.manager [-] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Took 1.03 seconds to deallocate network for instance. [ 668.228176] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Successfully created port: 9d80c545-2d7e-43eb-9773-6ed1e023fa91 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.234762] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Releasing lock "refresh_cache-b275344f-695e-463b-8bfe-2bee97296349" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.235216] env[63355]: DEBUG nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 668.235413] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 668.235703] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4774aa23-6472-4d61-a2f6-b648dcedb80b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.247824] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2dbbeb-344c-4946-8ab5-b17a2025b868 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.273745] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b275344f-695e-463b-8bfe-2bee97296349 could not be found. [ 668.274210] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 668.274440] env[63355]: INFO nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Took 0.04 seconds to destroy the instance on the hypervisor. [ 668.274735] env[63355]: DEBUG oslo.service.loopingcall [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 668.274972] env[63355]: DEBUG nova.compute.manager [-] [instance: b275344f-695e-463b-8bfe-2bee97296349] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.275083] env[63355]: DEBUG nova.network.neutron [-] [instance: b275344f-695e-463b-8bfe-2bee97296349] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 668.280409] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e2efc89a-f4a8-45ef-a98e-3b1ee1606bfb tempest-ServerExternalEventsTest-2035843949 tempest-ServerExternalEventsTest-2035843949-project-member] Lock "bc1d7402-3651-488b-a431-2cb2ae987d32" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.305s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.284543] env[63355]: DEBUG nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 668.309404] env[63355]: DEBUG nova.network.neutron [-] [instance: b275344f-695e-463b-8bfe-2bee97296349] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.626427] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquiring lock "a1e1e0ea-debd-4529-b175-4bdf44e4de31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.626954] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "a1e1e0ea-debd-4529-b175-4bdf44e4de31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.651704] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.710637] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "21e17d32-40db-4dc1-9e73-fa6d35a4c06a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.710637] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "21e17d32-40db-4dc1-9e73-fa6d35a4c06a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.780347] env[63355]: DEBUG nova.compute.manager [req-ea4c8ce5-24c7-4e85-885f-6fcfc7a9a346 req-a1ddc71c-7cdf-4476-8a59-0535c38a402b service nova] [instance: b275344f-695e-463b-8bfe-2bee97296349] Received event network-vif-deleted-c43c4079-a3a9-401e-ba17-aab8f7a656ad {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 668.783468] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 668.813025] env[63355]: DEBUG nova.network.neutron [-] [instance: b275344f-695e-463b-8bfe-2bee97296349] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.879707] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9bfd851-de5d-4a3c-b2d8-f5e219c3bd87 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.888963] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3291a3-a1e4-41ef-95b8-f49c709ab2f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.927806] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43256c16-1bf2-4fba-afcc-563ca4f5c968 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.935530] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85297b46-f091-4959-b347-5cc7d32af3e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.949398] env[63355]: DEBUG nova.compute.provider_tree [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.185269] env[63355]: ERROR nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9d80c545-2d7e-43eb-9773-6ed1e023fa91, please check neutron logs for more information. [ 669.185269] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.185269] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.185269] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.185269] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.185269] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.185269] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.185269] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.185269] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.185269] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 669.185269] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.185269] env[63355]: ERROR nova.compute.manager raise self.value [ 669.185269] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.185269] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.185269] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.185269] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.185692] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.185692] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.185692] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9d80c545-2d7e-43eb-9773-6ed1e023fa91, please check neutron logs for more information. [ 669.185692] env[63355]: ERROR nova.compute.manager [ 669.185692] env[63355]: Traceback (most recent call last): [ 669.185692] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.185692] env[63355]: listener.cb(fileno) [ 669.185692] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.185692] env[63355]: result = function(*args, **kwargs) [ 669.185692] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.185692] env[63355]: return func(*args, **kwargs) [ 669.185692] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.185692] env[63355]: raise e [ 669.185692] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.185692] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 669.185692] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.185692] env[63355]: created_port_ids = self._update_ports_for_instance( [ 669.185692] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.185692] env[63355]: with excutils.save_and_reraise_exception(): [ 669.185692] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.185692] env[63355]: self.force_reraise() [ 669.185692] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.185692] env[63355]: raise self.value [ 669.185692] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.185692] env[63355]: updated_port = self._update_port( [ 669.185692] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.185692] env[63355]: _ensure_no_port_binding_failure(port) [ 669.185692] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.185692] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.186488] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 9d80c545-2d7e-43eb-9773-6ed1e023fa91, please check neutron logs for more information. [ 669.186488] env[63355]: Removing descriptor: 17 [ 669.294803] env[63355]: DEBUG nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 669.316555] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.317884] env[63355]: INFO nova.compute.manager [-] [instance: b275344f-695e-463b-8bfe-2bee97296349] Took 1.04 seconds to deallocate network for instance. [ 669.321983] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.322218] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.322367] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.322538] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.322677] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.322814] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.323024] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.323184] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.323341] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.323520] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.323694] env[63355]: DEBUG nova.virt.hardware [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.324745] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90274f38-ba0c-428f-9eba-06b04b93f759 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.327872] env[63355]: DEBUG nova.compute.claims [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 669.328058] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.333634] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21491c72-cd82-43f1-bcbc-eb0acf6d4a6b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.349080] env[63355]: ERROR nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9d80c545-2d7e-43eb-9773-6ed1e023fa91, please check neutron logs for more information. [ 669.349080] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Traceback (most recent call last): [ 669.349080] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 669.349080] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] yield resources [ 669.349080] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.349080] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] self.driver.spawn(context, instance, image_meta, [ 669.349080] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.349080] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.349080] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.349080] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] vm_ref = self.build_virtual_machine(instance, [ 669.349080] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] for vif in network_info: [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] return self._sync_wrapper(fn, *args, **kwargs) [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] self.wait() [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] self[:] = self._gt.wait() [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] return self._exit_event.wait() [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 669.349377] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] current.throw(*self._exc) [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] result = function(*args, **kwargs) [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] return func(*args, **kwargs) [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] raise e [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] nwinfo = self.network_api.allocate_for_instance( [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] created_port_ids = self._update_ports_for_instance( [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] with excutils.save_and_reraise_exception(): [ 669.349686] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] self.force_reraise() [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] raise self.value [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] updated_port = self._update_port( [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] _ensure_no_port_binding_failure(port) [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] raise exception.PortBindingFailed(port_id=port['id']) [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] nova.exception.PortBindingFailed: Binding failed for port 9d80c545-2d7e-43eb-9773-6ed1e023fa91, please check neutron logs for more information. [ 669.349998] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] [ 669.349998] env[63355]: INFO nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Terminating instance [ 669.352307] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Acquiring lock "refresh_cache-3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.352516] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Acquired lock "refresh_cache-3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.353010] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.453566] env[63355]: DEBUG nova.scheduler.client.report [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.884606] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.959252] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.686s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.960193] env[63355]: DEBUG nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 669.966981] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.407s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.974187] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.224849] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquiring lock "7d79c9cf-66a9-4117-b090-0ab0676c3114" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.225140] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "7d79c9cf-66a9-4117-b090-0ab0676c3114" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.474981] env[63355]: DEBUG nova.compute.utils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 670.476383] env[63355]: DEBUG nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 670.478957] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 670.487158] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Releasing lock "refresh_cache-3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.488693] env[63355]: DEBUG nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 670.490464] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 670.492848] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1fa24fc-a0ae-4df8-bf31-69fee9d226a6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.506157] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93462d2d-4575-4726-b467-3c93e844afa9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.529254] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab could not be found. [ 670.529472] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 670.529654] env[63355]: INFO nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Took 0.04 seconds to destroy the instance on the hypervisor. [ 670.529894] env[63355]: DEBUG oslo.service.loopingcall [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 670.530123] env[63355]: DEBUG nova.compute.manager [-] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.530211] env[63355]: DEBUG nova.network.neutron [-] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.548680] env[63355]: DEBUG nova.policy [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'beb41983e69a44b2a108f552da5ef640', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfff1b9903264e5586119ebd3a3602de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 670.564780] env[63355]: DEBUG nova.network.neutron [-] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.957910] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Successfully created port: 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 670.991960] env[63355]: DEBUG nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 671.068404] env[63355]: DEBUG nova.network.neutron [-] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.094695] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7d7045-e950-4577-b84a-791d7e794f72 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.103980] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8330da-c013-4736-9b89-9bfb2215c115 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.142893] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0b746c-c115-4848-877a-26bde7aa7f31 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.151092] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad8b80a-510a-4186-9361-04220f855913 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.165721] env[63355]: DEBUG nova.compute.provider_tree [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.440347] env[63355]: DEBUG nova.compute.manager [req-426cf9c9-49ab-483c-9b16-4f462dbb26a1 req-cd30512d-941c-4d61-a2d0-bfa7c82ce389 service nova] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Received event network-changed-9d80c545-2d7e-43eb-9773-6ed1e023fa91 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.440599] env[63355]: DEBUG nova.compute.manager [req-426cf9c9-49ab-483c-9b16-4f462dbb26a1 req-cd30512d-941c-4d61-a2d0-bfa7c82ce389 service nova] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Refreshing instance network info cache due to event network-changed-9d80c545-2d7e-43eb-9773-6ed1e023fa91. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 671.440715] env[63355]: DEBUG oslo_concurrency.lockutils [req-426cf9c9-49ab-483c-9b16-4f462dbb26a1 req-cd30512d-941c-4d61-a2d0-bfa7c82ce389 service nova] Acquiring lock "refresh_cache-3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.440841] env[63355]: DEBUG oslo_concurrency.lockutils [req-426cf9c9-49ab-483c-9b16-4f462dbb26a1 req-cd30512d-941c-4d61-a2d0-bfa7c82ce389 service nova] Acquired lock "refresh_cache-3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.440993] env[63355]: DEBUG nova.network.neutron [req-426cf9c9-49ab-483c-9b16-4f462dbb26a1 req-cd30512d-941c-4d61-a2d0-bfa7c82ce389 service nova] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Refreshing network info cache for port 9d80c545-2d7e-43eb-9773-6ed1e023fa91 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 671.572089] env[63355]: INFO nova.compute.manager [-] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Took 1.04 seconds to deallocate network for instance. [ 671.574421] env[63355]: DEBUG nova.compute.claims [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 671.574642] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.668833] env[63355]: DEBUG nova.scheduler.client.report [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.963098] env[63355]: DEBUG nova.network.neutron [req-426cf9c9-49ab-483c-9b16-4f462dbb26a1 req-cd30512d-941c-4d61-a2d0-bfa7c82ce389 service nova] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.000238] env[63355]: DEBUG nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 672.044861] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 672.045190] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 672.045347] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 672.045529] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 672.045678] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 672.045829] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 672.046726] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 672.046726] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 672.046726] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 672.046726] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 672.046726] env[63355]: DEBUG nova.virt.hardware [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.047803] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ab84ae-9823-4cd9-9971-b23a7dcca439 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.056786] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb7de258-e472-41bc-8d3b-f3c0a76ed69b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.099246] env[63355]: DEBUG nova.network.neutron [req-426cf9c9-49ab-483c-9b16-4f462dbb26a1 req-cd30512d-941c-4d61-a2d0-bfa7c82ce389 service nova] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.174494] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.207s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.175494] env[63355]: ERROR nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 859440b2-5360-450e-9894-afa75c885899, please check neutron logs for more information. [ 672.175494] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Traceback (most recent call last): [ 672.175494] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.175494] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] self.driver.spawn(context, instance, image_meta, [ 672.175494] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 672.175494] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.175494] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.175494] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] vm_ref = self.build_virtual_machine(instance, [ 672.175494] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.175494] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.175494] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] for vif in network_info: [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] return self._sync_wrapper(fn, *args, **kwargs) [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] self.wait() [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] self[:] = self._gt.wait() [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] return self._exit_event.wait() [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] result = hub.switch() [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 672.176239] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] return self.greenlet.switch() [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] result = function(*args, **kwargs) [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] return func(*args, **kwargs) [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] raise e [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] nwinfo = self.network_api.allocate_for_instance( [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] created_port_ids = self._update_ports_for_instance( [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] with excutils.save_and_reraise_exception(): [ 672.176889] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] self.force_reraise() [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] raise self.value [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] updated_port = self._update_port( [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] _ensure_no_port_binding_failure(port) [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] raise exception.PortBindingFailed(port_id=port['id']) [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] nova.exception.PortBindingFailed: Binding failed for port 859440b2-5360-450e-9894-afa75c885899, please check neutron logs for more information. [ 672.177641] env[63355]: ERROR nova.compute.manager [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] [ 672.177956] env[63355]: DEBUG nova.compute.utils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Binding failed for port 859440b2-5360-450e-9894-afa75c885899, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 672.177989] env[63355]: DEBUG nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Build of instance 1de0e894-1edb-4296-9e9b-2c1eb0c7c275 was re-scheduled: Binding failed for port 859440b2-5360-450e-9894-afa75c885899, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 672.178431] env[63355]: DEBUG nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 672.178664] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Acquiring lock "refresh_cache-1de0e894-1edb-4296-9e9b-2c1eb0c7c275" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.178807] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Acquired lock "refresh_cache-1de0e894-1edb-4296-9e9b-2c1eb0c7c275" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.178960] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.179922] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.393s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.184093] env[63355]: INFO nova.compute.claims [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 672.269287] env[63355]: ERROR nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e, please check neutron logs for more information. [ 672.269287] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 672.269287] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.269287] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 672.269287] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.269287] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 672.269287] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.269287] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 672.269287] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.269287] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 672.269287] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.269287] env[63355]: ERROR nova.compute.manager raise self.value [ 672.269287] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.269287] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 672.269287] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.269287] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 672.269786] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.269786] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 672.269786] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e, please check neutron logs for more information. [ 672.269786] env[63355]: ERROR nova.compute.manager [ 672.269786] env[63355]: Traceback (most recent call last): [ 672.269786] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 672.269786] env[63355]: listener.cb(fileno) [ 672.269786] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.269786] env[63355]: result = function(*args, **kwargs) [ 672.269786] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.269786] env[63355]: return func(*args, **kwargs) [ 672.269786] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.269786] env[63355]: raise e [ 672.269786] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.269786] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 672.269786] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.269786] env[63355]: created_port_ids = self._update_ports_for_instance( [ 672.269786] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.269786] env[63355]: with excutils.save_and_reraise_exception(): [ 672.269786] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.269786] env[63355]: self.force_reraise() [ 672.269786] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.269786] env[63355]: raise self.value [ 672.269786] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.269786] env[63355]: updated_port = self._update_port( [ 672.269786] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.269786] env[63355]: _ensure_no_port_binding_failure(port) [ 672.269786] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.269786] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 672.270585] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e, please check neutron logs for more information. [ 672.270585] env[63355]: Removing descriptor: 17 [ 672.270585] env[63355]: ERROR nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e, please check neutron logs for more information. [ 672.270585] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Traceback (most recent call last): [ 672.270585] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 672.270585] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] yield resources [ 672.270585] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.270585] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] self.driver.spawn(context, instance, image_meta, [ 672.270585] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 672.270585] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.270585] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.270585] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] vm_ref = self.build_virtual_machine(instance, [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] for vif in network_info: [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] return self._sync_wrapper(fn, *args, **kwargs) [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] self.wait() [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] self[:] = self._gt.wait() [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] return self._exit_event.wait() [ 672.270963] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] result = hub.switch() [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] return self.greenlet.switch() [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] result = function(*args, **kwargs) [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] return func(*args, **kwargs) [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] raise e [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] nwinfo = self.network_api.allocate_for_instance( [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.271368] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] created_port_ids = self._update_ports_for_instance( [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] with excutils.save_and_reraise_exception(): [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] self.force_reraise() [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] raise self.value [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] updated_port = self._update_port( [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] _ensure_no_port_binding_failure(port) [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.272137] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] raise exception.PortBindingFailed(port_id=port['id']) [ 672.272551] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] nova.exception.PortBindingFailed: Binding failed for port 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e, please check neutron logs for more information. [ 672.272551] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] [ 672.272551] env[63355]: INFO nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Terminating instance [ 672.272551] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-0de6892a-5463-4d60-807c-0e3fca2accf3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.272551] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-0de6892a-5463-4d60-807c-0e3fca2accf3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.272703] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.601654] env[63355]: DEBUG oslo_concurrency.lockutils [req-426cf9c9-49ab-483c-9b16-4f462dbb26a1 req-cd30512d-941c-4d61-a2d0-bfa7c82ce389 service nova] Releasing lock "refresh_cache-3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.601927] env[63355]: DEBUG nova.compute.manager [req-426cf9c9-49ab-483c-9b16-4f462dbb26a1 req-cd30512d-941c-4d61-a2d0-bfa7c82ce389 service nova] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Received event network-vif-deleted-9d80c545-2d7e-43eb-9773-6ed1e023fa91 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 672.709967] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.790418] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.848142] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.950483] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.352547] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Releasing lock "refresh_cache-1de0e894-1edb-4296-9e9b-2c1eb0c7c275" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.354738] env[63355]: DEBUG nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 673.354738] env[63355]: DEBUG nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.354738] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.371785] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.456019] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-0de6892a-5463-4d60-807c-0e3fca2accf3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.456019] env[63355]: DEBUG nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 673.456019] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 673.456019] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c370ed61-04dd-4457-8105-40ef1f01c46e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.465192] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9af34e-f00e-4446-8a5a-f9a2179b9714 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.497929] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0de6892a-5463-4d60-807c-0e3fca2accf3 could not be found. [ 673.498237] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 673.498465] env[63355]: INFO nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 673.498755] env[63355]: DEBUG oslo.service.loopingcall [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.501358] env[63355]: DEBUG nova.compute.manager [-] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.501486] env[63355]: DEBUG nova.network.neutron [-] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.531075] env[63355]: DEBUG nova.network.neutron [-] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.655278] env[63355]: DEBUG nova.compute.manager [req-42274423-4096-44ef-8a9e-0bc443ac3c35 req-9cede6fd-4b74-4e16-863c-2e8a1053083f service nova] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Received event network-changed-71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 673.655278] env[63355]: DEBUG nova.compute.manager [req-42274423-4096-44ef-8a9e-0bc443ac3c35 req-9cede6fd-4b74-4e16-863c-2e8a1053083f service nova] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Refreshing instance network info cache due to event network-changed-71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 673.655278] env[63355]: DEBUG oslo_concurrency.lockutils [req-42274423-4096-44ef-8a9e-0bc443ac3c35 req-9cede6fd-4b74-4e16-863c-2e8a1053083f service nova] Acquiring lock "refresh_cache-0de6892a-5463-4d60-807c-0e3fca2accf3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.655278] env[63355]: DEBUG oslo_concurrency.lockutils [req-42274423-4096-44ef-8a9e-0bc443ac3c35 req-9cede6fd-4b74-4e16-863c-2e8a1053083f service nova] Acquired lock "refresh_cache-0de6892a-5463-4d60-807c-0e3fca2accf3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.655278] env[63355]: DEBUG nova.network.neutron [req-42274423-4096-44ef-8a9e-0bc443ac3c35 req-9cede6fd-4b74-4e16-863c-2e8a1053083f service nova] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Refreshing network info cache for port 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 673.773845] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d61bbeb-73e2-416a-9f5b-0f6d53518371 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.783167] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22b6277-9c12-4cc9-974a-c2c8cf385ffc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.818013] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47ef235-531f-4b19-b4d5-0ac9e3eb6749 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.825429] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2d6124-22fe-4ee0-b6a5-20a2bcb4e57d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.839716] env[63355]: DEBUG nova.compute.provider_tree [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.874531] env[63355]: DEBUG nova.network.neutron [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.036149] env[63355]: DEBUG nova.network.neutron [-] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.189821] env[63355]: DEBUG nova.network.neutron [req-42274423-4096-44ef-8a9e-0bc443ac3c35 req-9cede6fd-4b74-4e16-863c-2e8a1053083f service nova] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.322147] env[63355]: DEBUG nova.network.neutron [req-42274423-4096-44ef-8a9e-0bc443ac3c35 req-9cede6fd-4b74-4e16-863c-2e8a1053083f service nova] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.343344] env[63355]: DEBUG nova.scheduler.client.report [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.380017] env[63355]: INFO nova.compute.manager [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] [instance: 1de0e894-1edb-4296-9e9b-2c1eb0c7c275] Took 1.03 seconds to deallocate network for instance. [ 674.539781] env[63355]: INFO nova.compute.manager [-] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Took 1.04 seconds to deallocate network for instance. [ 674.541595] env[63355]: DEBUG nova.compute.claims [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 674.541792] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.824636] env[63355]: DEBUG oslo_concurrency.lockutils [req-42274423-4096-44ef-8a9e-0bc443ac3c35 req-9cede6fd-4b74-4e16-863c-2e8a1053083f service nova] Releasing lock "refresh_cache-0de6892a-5463-4d60-807c-0e3fca2accf3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.824942] env[63355]: DEBUG nova.compute.manager [req-42274423-4096-44ef-8a9e-0bc443ac3c35 req-9cede6fd-4b74-4e16-863c-2e8a1053083f service nova] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Received event network-vif-deleted-71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 674.850190] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.669s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.850190] env[63355]: DEBUG nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 674.855027] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.099s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.855027] env[63355]: INFO nova.compute.claims [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.359465] env[63355]: DEBUG nova.compute.utils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.364015] env[63355]: DEBUG nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.364015] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 675.422628] env[63355]: INFO nova.scheduler.client.report [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Deleted allocations for instance 1de0e894-1edb-4296-9e9b-2c1eb0c7c275 [ 675.488547] env[63355]: DEBUG nova.policy [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3721c4a73df54ae9b44110cb9e8590a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc547065748241e8ac7b6c499ddaea66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 675.870164] env[63355]: DEBUG nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 675.936229] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f5cdb38-ebcf-47c2-95f7-c67a46fdd161 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228 tempest-FloatingIPsAssociationNegativeTestJSON-2085345228-project-member] Lock "1de0e894-1edb-4296-9e9b-2c1eb0c7c275" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.585s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.203851] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Successfully created port: f34f3878-f7a3-41e9-84ee-8e31f5577cda {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.438751] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 676.456187] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0509199b-1cc6-4e15-b948-998135f458a6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.463638] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289d83ff-bf80-4971-b6ac-84706531bd24 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.495654] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4d4a3c-caba-438e-af74-5d68a8e92a17 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.503992] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cefd1473-7a4f-4a9a-bfde-8947d5dfce7d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.519749] env[63355]: DEBUG nova.compute.provider_tree [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.889188] env[63355]: DEBUG nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 676.930411] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 676.930411] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 676.930411] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 676.930411] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 676.930694] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 676.930694] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 676.930694] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 676.930694] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 676.930694] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 676.930819] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 676.930819] env[63355]: DEBUG nova.virt.hardware [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 676.932017] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056e30f1-de92-4265-8d55-cf5f6e637d47 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.940882] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4a4e75-f245-4c08-9f5e-d92c22980e57 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.977773] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.023128] env[63355]: DEBUG nova.scheduler.client.report [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.528484] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.676s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.529065] env[63355]: DEBUG nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 677.533782] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.672s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.534973] env[63355]: INFO nova.compute.claims [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 678.048546] env[63355]: DEBUG nova.compute.utils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.050306] env[63355]: DEBUG nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 678.050306] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 678.249973] env[63355]: ERROR nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f34f3878-f7a3-41e9-84ee-8e31f5577cda, please check neutron logs for more information. [ 678.249973] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 678.249973] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.249973] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 678.249973] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.249973] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 678.249973] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.249973] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 678.249973] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.249973] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 678.249973] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.249973] env[63355]: ERROR nova.compute.manager raise self.value [ 678.249973] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.249973] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 678.249973] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.249973] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 678.250596] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.250596] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 678.250596] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f34f3878-f7a3-41e9-84ee-8e31f5577cda, please check neutron logs for more information. [ 678.250596] env[63355]: ERROR nova.compute.manager [ 678.250596] env[63355]: Traceback (most recent call last): [ 678.250596] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 678.250596] env[63355]: listener.cb(fileno) [ 678.250596] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.250596] env[63355]: result = function(*args, **kwargs) [ 678.250596] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 678.250596] env[63355]: return func(*args, **kwargs) [ 678.250596] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.250596] env[63355]: raise e [ 678.250596] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.250596] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 678.250596] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.250596] env[63355]: created_port_ids = self._update_ports_for_instance( [ 678.250596] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.250596] env[63355]: with excutils.save_and_reraise_exception(): [ 678.250596] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.250596] env[63355]: self.force_reraise() [ 678.250596] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.250596] env[63355]: raise self.value [ 678.250596] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.250596] env[63355]: updated_port = self._update_port( [ 678.250596] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.250596] env[63355]: _ensure_no_port_binding_failure(port) [ 678.250596] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.250596] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 678.251400] env[63355]: nova.exception.PortBindingFailed: Binding failed for port f34f3878-f7a3-41e9-84ee-8e31f5577cda, please check neutron logs for more information. [ 678.251400] env[63355]: Removing descriptor: 18 [ 678.251400] env[63355]: ERROR nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f34f3878-f7a3-41e9-84ee-8e31f5577cda, please check neutron logs for more information. [ 678.251400] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Traceback (most recent call last): [ 678.251400] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 678.251400] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] yield resources [ 678.251400] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 678.251400] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] self.driver.spawn(context, instance, image_meta, [ 678.251400] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 678.251400] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] self._vmops.spawn(context, instance, image_meta, injected_files, [ 678.251400] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 678.251400] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] vm_ref = self.build_virtual_machine(instance, [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] vif_infos = vmwarevif.get_vif_info(self._session, [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] for vif in network_info: [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] return self._sync_wrapper(fn, *args, **kwargs) [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] self.wait() [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] self[:] = self._gt.wait() [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] return self._exit_event.wait() [ 678.251683] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] result = hub.switch() [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] return self.greenlet.switch() [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] result = function(*args, **kwargs) [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] return func(*args, **kwargs) [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] raise e [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] nwinfo = self.network_api.allocate_for_instance( [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.252014] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] created_port_ids = self._update_ports_for_instance( [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] with excutils.save_and_reraise_exception(): [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] self.force_reraise() [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] raise self.value [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] updated_port = self._update_port( [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] _ensure_no_port_binding_failure(port) [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.252328] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] raise exception.PortBindingFailed(port_id=port['id']) [ 678.252643] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] nova.exception.PortBindingFailed: Binding failed for port f34f3878-f7a3-41e9-84ee-8e31f5577cda, please check neutron logs for more information. [ 678.252643] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] [ 678.252643] env[63355]: INFO nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Terminating instance [ 678.255893] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "refresh_cache-2386393a-3006-4e6b-91e0-40b7ab8e3948" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.255893] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "refresh_cache-2386393a-3006-4e6b-91e0-40b7ab8e3948" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.255893] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.287521] env[63355]: DEBUG nova.policy [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe6b27b9050140f095dee9054ddee275', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '47cc7494ec7048eabe549433109dfda6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 678.387039] env[63355]: DEBUG nova.compute.manager [req-fcbb7f73-c089-4dac-86be-4edef2c17e30 req-f59480ed-f1af-4323-b0d6-f62f1640229a service nova] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Received event network-changed-f34f3878-f7a3-41e9-84ee-8e31f5577cda {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 678.387663] env[63355]: DEBUG nova.compute.manager [req-fcbb7f73-c089-4dac-86be-4edef2c17e30 req-f59480ed-f1af-4323-b0d6-f62f1640229a service nova] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Refreshing instance network info cache due to event network-changed-f34f3878-f7a3-41e9-84ee-8e31f5577cda. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 678.387663] env[63355]: DEBUG oslo_concurrency.lockutils [req-fcbb7f73-c089-4dac-86be-4edef2c17e30 req-f59480ed-f1af-4323-b0d6-f62f1640229a service nova] Acquiring lock "refresh_cache-2386393a-3006-4e6b-91e0-40b7ab8e3948" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.557723] env[63355]: DEBUG nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 678.743970] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Successfully created port: 1c6554f9-16ee-475e-a6cc-1c9e180c9ace {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 678.798346] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.031443] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquiring lock "4ded6053-9afc-444e-bdfc-244265ed19a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.031736] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "4ded6053-9afc-444e-bdfc-244265ed19a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.046228] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.161023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3260310-eacb-41dc-9da0-326bda8a5bc8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.173424] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6299b9-acb2-43d5-aad4-f70850013fe8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.202606] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb61224-a993-4dde-ab37-6588e96e122b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.210110] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11bce72-6f76-4265-9bce-5242e18eb891 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.223853] env[63355]: DEBUG nova.compute.provider_tree [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.550691] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "refresh_cache-2386393a-3006-4e6b-91e0-40b7ab8e3948" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.550691] env[63355]: DEBUG nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.550691] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 679.550691] env[63355]: DEBUG oslo_concurrency.lockutils [req-fcbb7f73-c089-4dac-86be-4edef2c17e30 req-f59480ed-f1af-4323-b0d6-f62f1640229a service nova] Acquired lock "refresh_cache-2386393a-3006-4e6b-91e0-40b7ab8e3948" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.551565] env[63355]: DEBUG nova.network.neutron [req-fcbb7f73-c089-4dac-86be-4edef2c17e30 req-f59480ed-f1af-4323-b0d6-f62f1640229a service nova] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Refreshing network info cache for port f34f3878-f7a3-41e9-84ee-8e31f5577cda {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 679.553504] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c4fcb19-c6ea-4359-9751-09805666542a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.568268] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd6e3cb-2360-491e-beb1-f564a41c00ca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.586717] env[63355]: DEBUG nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 679.602700] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2386393a-3006-4e6b-91e0-40b7ab8e3948 could not be found. [ 679.603368] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 679.603658] env[63355]: INFO nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Took 0.05 seconds to destroy the instance on the hypervisor. [ 679.604040] env[63355]: DEBUG oslo.service.loopingcall [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.604228] env[63355]: DEBUG nova.compute.manager [-] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.604350] env[63355]: DEBUG nova.network.neutron [-] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.616235] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 679.616481] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 679.616634] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.616813] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 679.616955] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.617127] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 679.617302] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 679.617451] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 679.617611] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 679.617770] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 679.617933] env[63355]: DEBUG nova.virt.hardware [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.618747] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2f1a1b-6670-426f-bfa2-ce3d3d009801 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.627337] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896ee5e8-a1ad-46e6-9ea7-1bdbf56315c9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.651178] env[63355]: DEBUG nova.network.neutron [-] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.729059] env[63355]: DEBUG nova.scheduler.client.report [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.057784] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquiring lock "dd80c3be-ec92-4a67-a373-3e2276752feb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.058114] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "dd80c3be-ec92-4a67-a373-3e2276752feb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.104207] env[63355]: DEBUG nova.network.neutron [req-fcbb7f73-c089-4dac-86be-4edef2c17e30 req-f59480ed-f1af-4323-b0d6-f62f1640229a service nova] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.119656] env[63355]: DEBUG nova.compute.manager [req-be0da03e-0ce1-4da8-a140-d60860f383e5 req-eed09fb1-9819-4646-b8f2-5b9b6d639fce service nova] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Received event network-changed-1c6554f9-16ee-475e-a6cc-1c9e180c9ace {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 680.119741] env[63355]: DEBUG nova.compute.manager [req-be0da03e-0ce1-4da8-a140-d60860f383e5 req-eed09fb1-9819-4646-b8f2-5b9b6d639fce service nova] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Refreshing instance network info cache due to event network-changed-1c6554f9-16ee-475e-a6cc-1c9e180c9ace. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 680.119941] env[63355]: DEBUG oslo_concurrency.lockutils [req-be0da03e-0ce1-4da8-a140-d60860f383e5 req-eed09fb1-9819-4646-b8f2-5b9b6d639fce service nova] Acquiring lock "refresh_cache-115e42ee-60cc-400d-8f4d-c23009f9091f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.120091] env[63355]: DEBUG oslo_concurrency.lockutils [req-be0da03e-0ce1-4da8-a140-d60860f383e5 req-eed09fb1-9819-4646-b8f2-5b9b6d639fce service nova] Acquired lock "refresh_cache-115e42ee-60cc-400d-8f4d-c23009f9091f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.120250] env[63355]: DEBUG nova.network.neutron [req-be0da03e-0ce1-4da8-a140-d60860f383e5 req-eed09fb1-9819-4646-b8f2-5b9b6d639fce service nova] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Refreshing network info cache for port 1c6554f9-16ee-475e-a6cc-1c9e180c9ace {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 680.131839] env[63355]: ERROR nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1c6554f9-16ee-475e-a6cc-1c9e180c9ace, please check neutron logs for more information. [ 680.131839] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.131839] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.131839] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.131839] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.131839] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.131839] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.131839] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.131839] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.131839] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 680.131839] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.131839] env[63355]: ERROR nova.compute.manager raise self.value [ 680.131839] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.131839] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.131839] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.131839] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.132314] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.132314] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.132314] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1c6554f9-16ee-475e-a6cc-1c9e180c9ace, please check neutron logs for more information. [ 680.132314] env[63355]: ERROR nova.compute.manager [ 680.132314] env[63355]: Traceback (most recent call last): [ 680.132314] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.132314] env[63355]: listener.cb(fileno) [ 680.132314] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.132314] env[63355]: result = function(*args, **kwargs) [ 680.132314] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.132314] env[63355]: return func(*args, **kwargs) [ 680.132314] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.132314] env[63355]: raise e [ 680.132314] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.132314] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 680.132314] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.132314] env[63355]: created_port_ids = self._update_ports_for_instance( [ 680.132314] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.132314] env[63355]: with excutils.save_and_reraise_exception(): [ 680.132314] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.132314] env[63355]: self.force_reraise() [ 680.132314] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.132314] env[63355]: raise self.value [ 680.132314] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.132314] env[63355]: updated_port = self._update_port( [ 680.132314] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.132314] env[63355]: _ensure_no_port_binding_failure(port) [ 680.132314] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.132314] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.133048] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 1c6554f9-16ee-475e-a6cc-1c9e180c9ace, please check neutron logs for more information. [ 680.133048] env[63355]: Removing descriptor: 17 [ 680.133048] env[63355]: ERROR nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1c6554f9-16ee-475e-a6cc-1c9e180c9ace, please check neutron logs for more information. [ 680.133048] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Traceback (most recent call last): [ 680.133048] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 680.133048] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] yield resources [ 680.133048] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.133048] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] self.driver.spawn(context, instance, image_meta, [ 680.133048] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.133048] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.133048] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.133048] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] vm_ref = self.build_virtual_machine(instance, [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] for vif in network_info: [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] return self._sync_wrapper(fn, *args, **kwargs) [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] self.wait() [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] self[:] = self._gt.wait() [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] return self._exit_event.wait() [ 680.133395] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] result = hub.switch() [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] return self.greenlet.switch() [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] result = function(*args, **kwargs) [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] return func(*args, **kwargs) [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] raise e [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] nwinfo = self.network_api.allocate_for_instance( [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.133779] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] created_port_ids = self._update_ports_for_instance( [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] with excutils.save_and_reraise_exception(): [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] self.force_reraise() [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] raise self.value [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] updated_port = self._update_port( [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] _ensure_no_port_binding_failure(port) [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.134114] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] raise exception.PortBindingFailed(port_id=port['id']) [ 680.134421] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] nova.exception.PortBindingFailed: Binding failed for port 1c6554f9-16ee-475e-a6cc-1c9e180c9ace, please check neutron logs for more information. [ 680.134421] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] [ 680.134421] env[63355]: INFO nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Terminating instance [ 680.135095] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquiring lock "refresh_cache-115e42ee-60cc-400d-8f4d-c23009f9091f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.155245] env[63355]: DEBUG nova.network.neutron [-] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.236194] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.702s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.236194] env[63355]: DEBUG nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 680.238771] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.825s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.271199] env[63355]: DEBUG nova.network.neutron [req-fcbb7f73-c089-4dac-86be-4edef2c17e30 req-f59480ed-f1af-4323-b0d6-f62f1640229a service nova] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.416866] env[63355]: DEBUG nova.compute.manager [req-c5ccc56e-5632-4af7-90ab-d4531a08e5bc req-711211ad-6883-4ead-822b-bd538ff46812 service nova] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Received event network-vif-deleted-f34f3878-f7a3-41e9-84ee-8e31f5577cda {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 680.637276] env[63355]: DEBUG nova.network.neutron [req-be0da03e-0ce1-4da8-a140-d60860f383e5 req-eed09fb1-9819-4646-b8f2-5b9b6d639fce service nova] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.658121] env[63355]: INFO nova.compute.manager [-] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Took 1.05 seconds to deallocate network for instance. [ 680.660411] env[63355]: DEBUG nova.compute.claims [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 680.660609] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.684433] env[63355]: DEBUG nova.network.neutron [req-be0da03e-0ce1-4da8-a140-d60860f383e5 req-eed09fb1-9819-4646-b8f2-5b9b6d639fce service nova] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.745524] env[63355]: DEBUG nova.compute.utils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 680.751814] env[63355]: DEBUG nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 680.752458] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 680.759690] env[63355]: DEBUG nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 680.772791] env[63355]: DEBUG oslo_concurrency.lockutils [req-fcbb7f73-c089-4dac-86be-4edef2c17e30 req-f59480ed-f1af-4323-b0d6-f62f1640229a service nova] Releasing lock "refresh_cache-2386393a-3006-4e6b-91e0-40b7ab8e3948" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.841429] env[63355]: DEBUG nova.policy [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcb346c7eb824f7e8f3fd47f3bbfa684', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e65f03e76d94d6aae439d7878a3c091', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 681.187804] env[63355]: DEBUG oslo_concurrency.lockutils [req-be0da03e-0ce1-4da8-a140-d60860f383e5 req-eed09fb1-9819-4646-b8f2-5b9b6d639fce service nova] Releasing lock "refresh_cache-115e42ee-60cc-400d-8f4d-c23009f9091f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.188752] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquired lock "refresh_cache-115e42ee-60cc-400d-8f4d-c23009f9091f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.189584] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 681.272361] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Successfully created port: 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 681.303669] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e490a3-61f1-47e0-983e-2e90e9cec4fe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.313144] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3165fc-8734-4f73-aa6b-fdf262d84b63 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.349297] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1fbc4ca-a6c2-43c3-9ce1-15d0cf688606 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.358476] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047544dd-8d31-4132-8dc7-5c59b726387b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.377109] env[63355]: DEBUG nova.compute.provider_tree [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.719191] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.771588] env[63355]: DEBUG nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 681.796498] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.802031] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.802163] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.805977] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.805977] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.805977] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.805977] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.805977] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.806438] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.806438] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.806438] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.806438] env[63355]: DEBUG nova.virt.hardware [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.806438] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3381d06d-3c73-49d9-8f9f-8cf77e7d3124 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.818023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255d734a-df25-4737-9655-87385b97baeb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.879929] env[63355]: DEBUG nova.scheduler.client.report [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.298347] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Releasing lock "refresh_cache-115e42ee-60cc-400d-8f4d-c23009f9091f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.298856] env[63355]: DEBUG nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 682.300431] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 682.301262] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa32e2b6-2847-436d-89d9-b00bd95710d3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.310319] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575f10e7-16c2-4a63-8e3e-a9c3293239be {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.324684] env[63355]: DEBUG nova.compute.manager [req-19434f3b-60f1-4467-8b19-3838015d6bf8 req-f5be0107-b0e9-43c9-a1f4-1808b9b7ca7d service nova] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Received event network-vif-deleted-1c6554f9-16ee-475e-a6cc-1c9e180c9ace {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.338484] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 115e42ee-60cc-400d-8f4d-c23009f9091f could not be found. [ 682.338734] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 682.338951] env[63355]: INFO nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 682.339236] env[63355]: DEBUG oslo.service.loopingcall [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 682.339470] env[63355]: DEBUG nova.compute.manager [-] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.339585] env[63355]: DEBUG nova.network.neutron [-] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 682.367568] env[63355]: DEBUG nova.network.neutron [-] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.389206] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.149s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.389206] env[63355]: ERROR nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 383c4b03-ec5b-4ef4-8b72-0dc09535c971, please check neutron logs for more information. [ 682.389206] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Traceback (most recent call last): [ 682.389206] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.389206] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] self.driver.spawn(context, instance, image_meta, [ 682.389206] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 682.389206] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.389206] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.389206] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] vm_ref = self.build_virtual_machine(instance, [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] for vif in network_info: [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] return self._sync_wrapper(fn, *args, **kwargs) [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] self.wait() [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] self[:] = self._gt.wait() [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] return self._exit_event.wait() [ 682.389608] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] result = hub.switch() [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] return self.greenlet.switch() [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] result = function(*args, **kwargs) [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] return func(*args, **kwargs) [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] raise e [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] nwinfo = self.network_api.allocate_for_instance( [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.390246] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] created_port_ids = self._update_ports_for_instance( [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] with excutils.save_and_reraise_exception(): [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] self.force_reraise() [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] raise self.value [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] updated_port = self._update_port( [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] _ensure_no_port_binding_failure(port) [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.390836] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] raise exception.PortBindingFailed(port_id=port['id']) [ 682.391408] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] nova.exception.PortBindingFailed: Binding failed for port 383c4b03-ec5b-4ef4-8b72-0dc09535c971, please check neutron logs for more information. [ 682.391408] env[63355]: ERROR nova.compute.manager [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] [ 682.391408] env[63355]: DEBUG nova.compute.utils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Binding failed for port 383c4b03-ec5b-4ef4-8b72-0dc09535c971, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.391408] env[63355]: DEBUG oslo_concurrency.lockutils [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.831s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.391408] env[63355]: DEBUG nova.objects.instance [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63355) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 682.393907] env[63355]: DEBUG nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Build of instance 10112b62-f41e-4413-9bc8-c1306c5c1bc1 was re-scheduled: Binding failed for port 383c4b03-ec5b-4ef4-8b72-0dc09535c971, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 682.395195] env[63355]: DEBUG nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 682.395438] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Acquiring lock "refresh_cache-10112b62-f41e-4413-9bc8-c1306c5c1bc1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.395584] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Acquired lock "refresh_cache-10112b62-f41e-4413-9bc8-c1306c5c1bc1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.395738] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.707235] env[63355]: DEBUG nova.compute.manager [req-292e5eab-04a7-45b9-a9c9-33d62b01d6a3 req-180e6ada-f516-4092-9109-b200b13dff38 service nova] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Received event network-changed-905828d8-0121-4c4e-83e0-cb7e6e4e6d2c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.707425] env[63355]: DEBUG nova.compute.manager [req-292e5eab-04a7-45b9-a9c9-33d62b01d6a3 req-180e6ada-f516-4092-9109-b200b13dff38 service nova] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Refreshing instance network info cache due to event network-changed-905828d8-0121-4c4e-83e0-cb7e6e4e6d2c. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 682.707629] env[63355]: DEBUG oslo_concurrency.lockutils [req-292e5eab-04a7-45b9-a9c9-33d62b01d6a3 req-180e6ada-f516-4092-9109-b200b13dff38 service nova] Acquiring lock "refresh_cache-b46098f3-d952-4a33-86fe-825fe4625008" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.707949] env[63355]: DEBUG oslo_concurrency.lockutils [req-292e5eab-04a7-45b9-a9c9-33d62b01d6a3 req-180e6ada-f516-4092-9109-b200b13dff38 service nova] Acquired lock "refresh_cache-b46098f3-d952-4a33-86fe-825fe4625008" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.707949] env[63355]: DEBUG nova.network.neutron [req-292e5eab-04a7-45b9-a9c9-33d62b01d6a3 req-180e6ada-f516-4092-9109-b200b13dff38 service nova] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Refreshing network info cache for port 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 682.856112] env[63355]: ERROR nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c, please check neutron logs for more information. [ 682.856112] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 682.856112] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.856112] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 682.856112] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.856112] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 682.856112] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.856112] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 682.856112] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.856112] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 682.856112] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.856112] env[63355]: ERROR nova.compute.manager raise self.value [ 682.856112] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.856112] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 682.856112] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.856112] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 682.856750] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.856750] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 682.856750] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c, please check neutron logs for more information. [ 682.856750] env[63355]: ERROR nova.compute.manager [ 682.856750] env[63355]: Traceback (most recent call last): [ 682.856750] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 682.856750] env[63355]: listener.cb(fileno) [ 682.856750] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.856750] env[63355]: result = function(*args, **kwargs) [ 682.856750] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.856750] env[63355]: return func(*args, **kwargs) [ 682.856750] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.856750] env[63355]: raise e [ 682.856750] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.856750] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 682.856750] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.856750] env[63355]: created_port_ids = self._update_ports_for_instance( [ 682.856750] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.856750] env[63355]: with excutils.save_and_reraise_exception(): [ 682.856750] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.856750] env[63355]: self.force_reraise() [ 682.856750] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.856750] env[63355]: raise self.value [ 682.856750] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.856750] env[63355]: updated_port = self._update_port( [ 682.856750] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.856750] env[63355]: _ensure_no_port_binding_failure(port) [ 682.856750] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.856750] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 682.857409] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c, please check neutron logs for more information. [ 682.857409] env[63355]: Removing descriptor: 18 [ 682.857950] env[63355]: ERROR nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c, please check neutron logs for more information. [ 682.857950] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] Traceback (most recent call last): [ 682.857950] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 682.857950] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] yield resources [ 682.857950] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.857950] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] self.driver.spawn(context, instance, image_meta, [ 682.857950] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 682.857950] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.857950] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.857950] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] vm_ref = self.build_virtual_machine(instance, [ 682.857950] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] for vif in network_info: [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] return self._sync_wrapper(fn, *args, **kwargs) [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] self.wait() [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] self[:] = self._gt.wait() [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] return self._exit_event.wait() [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 682.858228] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] result = hub.switch() [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] return self.greenlet.switch() [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] result = function(*args, **kwargs) [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] return func(*args, **kwargs) [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] raise e [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] nwinfo = self.network_api.allocate_for_instance( [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] created_port_ids = self._update_ports_for_instance( [ 682.858671] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] with excutils.save_and_reraise_exception(): [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] self.force_reraise() [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] raise self.value [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] updated_port = self._update_port( [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] _ensure_no_port_binding_failure(port) [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] raise exception.PortBindingFailed(port_id=port['id']) [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] nova.exception.PortBindingFailed: Binding failed for port 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c, please check neutron logs for more information. [ 682.859687] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] [ 682.860248] env[63355]: INFO nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Terminating instance [ 682.860697] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Acquiring lock "refresh_cache-b46098f3-d952-4a33-86fe-825fe4625008" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.874410] env[63355]: DEBUG nova.network.neutron [-] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.920839] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.020832] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.247016] env[63355]: DEBUG nova.network.neutron [req-292e5eab-04a7-45b9-a9c9-33d62b01d6a3 req-180e6ada-f516-4092-9109-b200b13dff38 service nova] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.379257] env[63355]: INFO nova.compute.manager [-] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Took 1.04 seconds to deallocate network for instance. [ 683.379936] env[63355]: DEBUG nova.compute.claims [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 683.380273] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.389325] env[63355]: DEBUG nova.network.neutron [req-292e5eab-04a7-45b9-a9c9-33d62b01d6a3 req-180e6ada-f516-4092-9109-b200b13dff38 service nova] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.403031] env[63355]: DEBUG oslo_concurrency.lockutils [None req-acc9a73a-9614-4ebd-bbdb-b11dc920a39d tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.403753] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.823s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.405672] env[63355]: INFO nova.compute.claims [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.036526] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Releasing lock "refresh_cache-10112b62-f41e-4413-9bc8-c1306c5c1bc1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.036854] env[63355]: DEBUG nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 684.036919] env[63355]: DEBUG nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.039153] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 684.039153] env[63355]: DEBUG oslo_concurrency.lockutils [req-292e5eab-04a7-45b9-a9c9-33d62b01d6a3 req-180e6ada-f516-4092-9109-b200b13dff38 service nova] Releasing lock "refresh_cache-b46098f3-d952-4a33-86fe-825fe4625008" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.041671] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Acquired lock "refresh_cache-b46098f3-d952-4a33-86fe-825fe4625008" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.041856] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 684.062161] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.564662] env[63355]: DEBUG nova.network.neutron [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.567784] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.702300] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.901379] env[63355]: DEBUG nova.compute.manager [req-00522057-8280-480f-a8e0-b1084723eef3 req-d38ebc55-df4d-4f4c-a41d-beb7dc547153 service nova] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Received event network-vif-deleted-905828d8-0121-4c4e-83e0-cb7e6e4e6d2c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 685.073172] env[63355]: INFO nova.compute.manager [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] [instance: 10112b62-f41e-4413-9bc8-c1306c5c1bc1] Took 1.03 seconds to deallocate network for instance. [ 685.122046] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf5c616-8b3d-4987-9aec-5763bb8d0e62 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.134111] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d301b565-b840-4599-9934-f912e3d6c08c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.170860] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dcaa4ea-cb21-44d6-ae7a-3808df75defa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.179678] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ad6e18-2ef4-4d2b-bab4-74ed278304bf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.194101] env[63355]: DEBUG nova.compute.provider_tree [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.205363] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Releasing lock "refresh_cache-b46098f3-d952-4a33-86fe-825fe4625008" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.205766] env[63355]: DEBUG nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 685.205998] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 685.206809] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-caa9475d-0502-46a5-b8d8-03116b9fff6a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.219672] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3247d2-744e-42b8-a97c-680ae5eb7d5d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.241620] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b46098f3-d952-4a33-86fe-825fe4625008 could not be found. [ 685.241825] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 685.242013] env[63355]: INFO nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Took 0.04 seconds to destroy the instance on the hypervisor. [ 685.242259] env[63355]: DEBUG oslo.service.loopingcall [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 685.243111] env[63355]: DEBUG nova.compute.manager [-] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.243111] env[63355]: DEBUG nova.network.neutron [-] [instance: b46098f3-d952-4a33-86fe-825fe4625008] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.264345] env[63355]: DEBUG nova.network.neutron [-] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.697941] env[63355]: DEBUG nova.scheduler.client.report [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.766924] env[63355]: DEBUG nova.network.neutron [-] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.108653] env[63355]: INFO nova.scheduler.client.report [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Deleted allocations for instance 10112b62-f41e-4413-9bc8-c1306c5c1bc1 [ 686.202618] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.799s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.203179] env[63355]: DEBUG nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 686.206957] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.556s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.207195] env[63355]: DEBUG nova.objects.instance [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lazy-loading 'resources' on Instance uuid f9c8ed86-506b-4654-8e7e-cb218605764b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 686.268761] env[63355]: INFO nova.compute.manager [-] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Took 1.03 seconds to deallocate network for instance. [ 686.271199] env[63355]: DEBUG nova.compute.claims [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 686.271407] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.616872] env[63355]: DEBUG oslo_concurrency.lockutils [None req-85783341-c706-4336-ae2a-d6cd25974722 tempest-ServerDiagnosticsNegativeTest-1847785470 tempest-ServerDiagnosticsNegativeTest-1847785470-project-member] Lock "10112b62-f41e-4413-9bc8-c1306c5c1bc1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.495s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.710416] env[63355]: DEBUG nova.compute.utils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 686.713982] env[63355]: DEBUG nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 686.714177] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 686.796667] env[63355]: DEBUG nova.policy [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77243643ea724b72858a8682a2a054a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d47b671ea9c429391cbdae7e24adadf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.119335] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.209281] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c5a786-9526-4063-b87d-5e4f35bf11f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.218639] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d226d8f-993c-4a05-b557-efff24881cfb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.228053] env[63355]: DEBUG nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 687.262745] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28768582-0221-4cab-8f25-2c38ba969cd7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.272986] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64dbec5e-a5cb-4aad-a980-194f90eca794 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.285389] env[63355]: DEBUG nova.compute.provider_tree [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.287258] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Successfully created port: b75086db-4824-445c-8b78-912828bec10e {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.640988] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.791559] env[63355]: DEBUG nova.scheduler.client.report [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.247727] env[63355]: DEBUG nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 688.285599] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.285599] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.285599] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.285845] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.286059] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.286200] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.286473] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.286711] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.287134] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.287357] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.287565] env[63355]: DEBUG nova.virt.hardware [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.288469] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1408a2e0-ddeb-4f3d-a267-e68acd699f87 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.297809] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.091s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.301626] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.985s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.302729] env[63355]: INFO nova.compute.claims [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 688.306861] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15f2217-0dd3-4399-8c39-2e76ffd35597 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.335753] env[63355]: INFO nova.scheduler.client.report [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleted allocations for instance f9c8ed86-506b-4654-8e7e-cb218605764b [ 688.844585] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbe38457-1742-41f6-b504-9f9e860a5f02 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "f9c8ed86-506b-4654-8e7e-cb218605764b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.969s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.960024] env[63355]: DEBUG nova.compute.manager [req-8030048f-85b0-4d1a-9b17-91b387186c69 req-859823f3-2c34-4f11-8f23-0027ef391bd9 service nova] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Received event network-changed-b75086db-4824-445c-8b78-912828bec10e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 688.960261] env[63355]: DEBUG nova.compute.manager [req-8030048f-85b0-4d1a-9b17-91b387186c69 req-859823f3-2c34-4f11-8f23-0027ef391bd9 service nova] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Refreshing instance network info cache due to event network-changed-b75086db-4824-445c-8b78-912828bec10e. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 688.960445] env[63355]: DEBUG oslo_concurrency.lockutils [req-8030048f-85b0-4d1a-9b17-91b387186c69 req-859823f3-2c34-4f11-8f23-0027ef391bd9 service nova] Acquiring lock "refresh_cache-732ff6d3-2de9-4e6b-887d-5329ba83443c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.960604] env[63355]: DEBUG oslo_concurrency.lockutils [req-8030048f-85b0-4d1a-9b17-91b387186c69 req-859823f3-2c34-4f11-8f23-0027ef391bd9 service nova] Acquired lock "refresh_cache-732ff6d3-2de9-4e6b-887d-5329ba83443c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.960758] env[63355]: DEBUG nova.network.neutron [req-8030048f-85b0-4d1a-9b17-91b387186c69 req-859823f3-2c34-4f11-8f23-0027ef391bd9 service nova] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Refreshing network info cache for port b75086db-4824-445c-8b78-912828bec10e {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 689.054854] env[63355]: ERROR nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b75086db-4824-445c-8b78-912828bec10e, please check neutron logs for more information. [ 689.054854] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.054854] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.054854] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.054854] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.054854] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.054854] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.054854] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.054854] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.054854] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 689.054854] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.054854] env[63355]: ERROR nova.compute.manager raise self.value [ 689.054854] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.054854] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.054854] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.054854] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.055433] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.055433] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.055433] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b75086db-4824-445c-8b78-912828bec10e, please check neutron logs for more information. [ 689.055433] env[63355]: ERROR nova.compute.manager [ 689.055433] env[63355]: Traceback (most recent call last): [ 689.055433] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.055433] env[63355]: listener.cb(fileno) [ 689.055433] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.055433] env[63355]: result = function(*args, **kwargs) [ 689.055433] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.055433] env[63355]: return func(*args, **kwargs) [ 689.055433] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.055433] env[63355]: raise e [ 689.055433] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.055433] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 689.055433] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.055433] env[63355]: created_port_ids = self._update_ports_for_instance( [ 689.055433] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.055433] env[63355]: with excutils.save_and_reraise_exception(): [ 689.055433] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.055433] env[63355]: self.force_reraise() [ 689.055433] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.055433] env[63355]: raise self.value [ 689.055433] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.055433] env[63355]: updated_port = self._update_port( [ 689.055433] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.055433] env[63355]: _ensure_no_port_binding_failure(port) [ 689.055433] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.055433] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.056135] env[63355]: nova.exception.PortBindingFailed: Binding failed for port b75086db-4824-445c-8b78-912828bec10e, please check neutron logs for more information. [ 689.056135] env[63355]: Removing descriptor: 18 [ 689.056135] env[63355]: ERROR nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b75086db-4824-445c-8b78-912828bec10e, please check neutron logs for more information. [ 689.056135] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Traceback (most recent call last): [ 689.056135] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 689.056135] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] yield resources [ 689.056135] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.056135] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] self.driver.spawn(context, instance, image_meta, [ 689.056135] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.056135] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.056135] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.056135] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] vm_ref = self.build_virtual_machine(instance, [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] for vif in network_info: [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] return self._sync_wrapper(fn, *args, **kwargs) [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] self.wait() [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] self[:] = self._gt.wait() [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] return self._exit_event.wait() [ 689.056477] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] result = hub.switch() [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] return self.greenlet.switch() [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] result = function(*args, **kwargs) [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] return func(*args, **kwargs) [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] raise e [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] nwinfo = self.network_api.allocate_for_instance( [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.056786] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] created_port_ids = self._update_ports_for_instance( [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] with excutils.save_and_reraise_exception(): [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] self.force_reraise() [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] raise self.value [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] updated_port = self._update_port( [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] _ensure_no_port_binding_failure(port) [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.057101] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] raise exception.PortBindingFailed(port_id=port['id']) [ 689.057395] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] nova.exception.PortBindingFailed: Binding failed for port b75086db-4824-445c-8b78-912828bec10e, please check neutron logs for more information. [ 689.057395] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] [ 689.057395] env[63355]: INFO nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Terminating instance [ 689.058288] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-732ff6d3-2de9-4e6b-887d-5329ba83443c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.501016] env[63355]: DEBUG nova.network.neutron [req-8030048f-85b0-4d1a-9b17-91b387186c69 req-859823f3-2c34-4f11-8f23-0027ef391bd9 service nova] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.658336] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "def90353-ae90-4e1e-9fe1-8e5459a7309d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.658633] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "def90353-ae90-4e1e-9fe1-8e5459a7309d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.658895] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "def90353-ae90-4e1e-9fe1-8e5459a7309d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.659154] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "def90353-ae90-4e1e-9fe1-8e5459a7309d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.659295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "def90353-ae90-4e1e-9fe1-8e5459a7309d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.664025] env[63355]: INFO nova.compute.manager [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Terminating instance [ 689.668159] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "refresh_cache-def90353-ae90-4e1e-9fe1-8e5459a7309d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.668439] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquired lock "refresh_cache-def90353-ae90-4e1e-9fe1-8e5459a7309d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.668516] env[63355]: DEBUG nova.network.neutron [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.723016] env[63355]: DEBUG nova.network.neutron [req-8030048f-85b0-4d1a-9b17-91b387186c69 req-859823f3-2c34-4f11-8f23-0027ef391bd9 service nova] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.826310] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20553d80-d8f3-4da5-b14d-bad6c87b3e22 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.835549] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f646235-640c-4f02-92fb-e78af4ccd33e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.867755] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309c09de-f5df-48d7-bdc4-df23762da022 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.875919] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9176384-7be0-4d35-a732-d51154395d48 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.890280] env[63355]: DEBUG nova.compute.provider_tree [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.211110] env[63355]: DEBUG nova.network.neutron [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.228620] env[63355]: DEBUG oslo_concurrency.lockutils [req-8030048f-85b0-4d1a-9b17-91b387186c69 req-859823f3-2c34-4f11-8f23-0027ef391bd9 service nova] Releasing lock "refresh_cache-732ff6d3-2de9-4e6b-887d-5329ba83443c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.229203] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-732ff6d3-2de9-4e6b-887d-5329ba83443c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.229287] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.393660] env[63355]: DEBUG nova.scheduler.client.report [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.484191] env[63355]: DEBUG nova.network.neutron [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.564155] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "9989c129-07d3-4af9-9a86-1e2746d6ac70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.565499] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "9989c129-07d3-4af9-9a86-1e2746d6ac70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.765277] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.895445] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.900767] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.598s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.900767] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.901155] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.573s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.986426] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Releasing lock "refresh_cache-def90353-ae90-4e1e-9fe1-8e5459a7309d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.986923] env[63355]: DEBUG nova.compute.manager [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 690.987319] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 690.988151] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d01821-22e0-4a03-b977-eade07641bac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.998596] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 690.999174] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3511846d-233e-4c57-b4b7-71e66db0bd7d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.007423] env[63355]: DEBUG oslo_vmware.api [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 691.007423] env[63355]: value = "task-1349518" [ 691.007423] env[63355]: _type = "Task" [ 691.007423] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.015215] env[63355]: DEBUG oslo_vmware.api [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349518, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.151602] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 691.152321] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 691.218679] env[63355]: DEBUG nova.compute.manager [req-78ac4fce-6bfe-4b3f-8b2d-9212cf0b71ee req-3e6f55b2-7fa6-4491-a5d5-155d273fe1b2 service nova] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Received event network-vif-deleted-b75086db-4824-445c-8b78-912828bec10e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.402341] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-732ff6d3-2de9-4e6b-887d-5329ba83443c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.402810] env[63355]: DEBUG nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 691.403182] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 691.403395] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ffa09d1-9079-43fe-a83d-24b65d7ef610 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.413654] env[63355]: DEBUG nova.compute.utils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 691.418458] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d24c859-8aef-42c4-adce-fa6c50f57647 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.430561] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.432982] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 691.433166] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 691.450892] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 732ff6d3-2de9-4e6b-887d-5329ba83443c could not be found. [ 691.452207] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 691.452207] env[63355]: INFO nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 691.452207] env[63355]: DEBUG oslo.service.loopingcall [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.452207] env[63355]: DEBUG nova.compute.manager [-] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.452380] env[63355]: DEBUG nova.network.neutron [-] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.469337] env[63355]: DEBUG nova.network.neutron [-] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.508057] env[63355]: DEBUG nova.policy [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b81f53d014ce4e5d937156d9a3e32c0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2bfdd9344794c6987e9915cf414b8d6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 691.520963] env[63355]: DEBUG oslo_vmware.api [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349518, 'name': PowerOffVM_Task, 'duration_secs': 0.134049} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.521252] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 691.521420] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 691.521908] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b0f04fa-db88-4265-ab4c-12600060e4d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.551196] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 691.551393] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 691.551712] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleting the datastore file [datastore1] def90353-ae90-4e1e-9fe1-8e5459a7309d {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 691.551953] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d08925c2-bf2d-46fa-acd6-02fe72643393 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.558610] env[63355]: DEBUG oslo_vmware.api [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for the task: (returnval){ [ 691.558610] env[63355]: value = "task-1349520" [ 691.558610] env[63355]: _type = "Task" [ 691.558610] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.570438] env[63355]: DEBUG oslo_vmware.api [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349520, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.574628] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "8a44bcfb-406c-45f6-b027-86d8980bc5d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.574958] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "8a44bcfb-406c-45f6-b027-86d8980bc5d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.663910] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 691.663910] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 691.663910] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Rebuilding the list of instances to heal {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 691.863584] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Successfully created port: 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.930086] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b39caff-dfe8-44fb-ade6-1edd21ca1421 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.939033] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874c37ab-70d1-42ff-a035-fdc612520c70 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.979239] env[63355]: DEBUG nova.network.neutron [-] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.981317] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceddc831-dedf-4698-b20e-533e1f7c3f09 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.989836] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f9f91b-3dad-4ddf-b65e-d897257a5285 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.004343] env[63355]: DEBUG nova.compute.provider_tree [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.068060] env[63355]: DEBUG oslo_vmware.api [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Task: {'id': task-1349520, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091385} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.068664] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 692.068664] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 692.068664] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 692.068798] env[63355]: INFO nova.compute.manager [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Took 1.08 seconds to destroy the instance on the hypervisor. [ 692.069038] env[63355]: DEBUG oslo.service.loopingcall [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 692.069217] env[63355]: DEBUG nova.compute.manager [-] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 692.069307] env[63355]: DEBUG nova.network.neutron [-] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 692.088859] env[63355]: DEBUG nova.network.neutron [-] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.169592] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Skipping network cache update for instance because it is being deleted. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 692.169983] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 692.169983] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 692.170138] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 692.170138] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 692.170254] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 692.170374] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 692.170491] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 692.170610] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Didn't find any instances for network info cache update. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 692.170801] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 692.170953] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 692.171405] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 692.171693] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 692.171891] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 692.172091] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 692.172256] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 692.172447] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 692.450346] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.488580] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.488580] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.488580] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.488942] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.488942] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.488942] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.488942] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.488942] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.489130] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.489130] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.489130] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.489130] env[63355]: INFO nova.compute.manager [-] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Took 1.04 seconds to deallocate network for instance. [ 692.489130] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e90f71-c005-41a9-9441-671d0aa4b926 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.493101] env[63355]: DEBUG nova.compute.claims [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 692.493301] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.497608] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aafcde73-ca87-4d88-8f9d-78eeb52351c5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.510983] env[63355]: DEBUG nova.scheduler.client.report [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.591441] env[63355]: DEBUG nova.network.neutron [-] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.678081] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.016397] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.114s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.016397] env[63355]: ERROR nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c43c4079-a3a9-401e-ba17-aab8f7a656ad, please check neutron logs for more information. [ 693.016397] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] Traceback (most recent call last): [ 693.016397] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.016397] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] self.driver.spawn(context, instance, image_meta, [ 693.016397] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 693.016397] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.016397] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.016397] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] vm_ref = self.build_virtual_machine(instance, [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] for vif in network_info: [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] return self._sync_wrapper(fn, *args, **kwargs) [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] self.wait() [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] self[:] = self._gt.wait() [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] return self._exit_event.wait() [ 693.017293] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] result = hub.switch() [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] return self.greenlet.switch() [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] result = function(*args, **kwargs) [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] return func(*args, **kwargs) [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] raise e [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] nwinfo = self.network_api.allocate_for_instance( [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.018069] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] created_port_ids = self._update_ports_for_instance( [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] with excutils.save_and_reraise_exception(): [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] self.force_reraise() [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] raise self.value [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] updated_port = self._update_port( [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] _ensure_no_port_binding_failure(port) [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.018445] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] raise exception.PortBindingFailed(port_id=port['id']) [ 693.018770] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] nova.exception.PortBindingFailed: Binding failed for port c43c4079-a3a9-401e-ba17-aab8f7a656ad, please check neutron logs for more information. [ 693.018770] env[63355]: ERROR nova.compute.manager [instance: b275344f-695e-463b-8bfe-2bee97296349] [ 693.018770] env[63355]: DEBUG nova.compute.utils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Binding failed for port c43c4079-a3a9-401e-ba17-aab8f7a656ad, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 693.018770] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.444s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.021304] env[63355]: DEBUG nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Build of instance b275344f-695e-463b-8bfe-2bee97296349 was re-scheduled: Binding failed for port c43c4079-a3a9-401e-ba17-aab8f7a656ad, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 693.021650] env[63355]: DEBUG nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 693.022513] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Acquiring lock "refresh_cache-b275344f-695e-463b-8bfe-2bee97296349" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.022513] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Acquired lock "refresh_cache-b275344f-695e-463b-8bfe-2bee97296349" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.022513] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.095130] env[63355]: INFO nova.compute.manager [-] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Took 1.02 seconds to deallocate network for instance. [ 693.168888] env[63355]: ERROR nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd, please check neutron logs for more information. [ 693.168888] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 693.168888] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.168888] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 693.168888] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.168888] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 693.168888] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.168888] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 693.168888] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.168888] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 693.168888] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.168888] env[63355]: ERROR nova.compute.manager raise self.value [ 693.168888] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.168888] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 693.168888] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.168888] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 693.169463] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.169463] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 693.169463] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd, please check neutron logs for more information. [ 693.169463] env[63355]: ERROR nova.compute.manager [ 693.169463] env[63355]: Traceback (most recent call last): [ 693.169463] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 693.169463] env[63355]: listener.cb(fileno) [ 693.169463] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.169463] env[63355]: result = function(*args, **kwargs) [ 693.169463] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 693.169463] env[63355]: return func(*args, **kwargs) [ 693.169463] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.169463] env[63355]: raise e [ 693.169463] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.169463] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 693.169463] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.169463] env[63355]: created_port_ids = self._update_ports_for_instance( [ 693.169463] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.169463] env[63355]: with excutils.save_and_reraise_exception(): [ 693.169463] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.169463] env[63355]: self.force_reraise() [ 693.169463] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.169463] env[63355]: raise self.value [ 693.169463] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.169463] env[63355]: updated_port = self._update_port( [ 693.169463] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.169463] env[63355]: _ensure_no_port_binding_failure(port) [ 693.169463] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.169463] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 693.170359] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd, please check neutron logs for more information. [ 693.170359] env[63355]: Removing descriptor: 18 [ 693.170359] env[63355]: ERROR nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd, please check neutron logs for more information. [ 693.170359] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Traceback (most recent call last): [ 693.170359] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 693.170359] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] yield resources [ 693.170359] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.170359] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] self.driver.spawn(context, instance, image_meta, [ 693.170359] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 693.170359] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.170359] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.170359] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] vm_ref = self.build_virtual_machine(instance, [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] for vif in network_info: [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] return self._sync_wrapper(fn, *args, **kwargs) [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] self.wait() [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] self[:] = self._gt.wait() [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] return self._exit_event.wait() [ 693.170707] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] result = hub.switch() [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] return self.greenlet.switch() [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] result = function(*args, **kwargs) [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] return func(*args, **kwargs) [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] raise e [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] nwinfo = self.network_api.allocate_for_instance( [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.171073] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] created_port_ids = self._update_ports_for_instance( [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] with excutils.save_and_reraise_exception(): [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] self.force_reraise() [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] raise self.value [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] updated_port = self._update_port( [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] _ensure_no_port_binding_failure(port) [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.171444] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] raise exception.PortBindingFailed(port_id=port['id']) [ 693.171768] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] nova.exception.PortBindingFailed: Binding failed for port 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd, please check neutron logs for more information. [ 693.171768] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] [ 693.171768] env[63355]: INFO nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Terminating instance [ 693.172397] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "refresh_cache-087c0c34-2f44-4791-86e7-b1b3564f49d5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.172552] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquired lock "refresh_cache-087c0c34-2f44-4791-86e7-b1b3564f49d5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.172720] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.340582] env[63355]: DEBUG nova.compute.manager [req-ae535077-6f45-42b6-9e2d-d43f8269f9ba req-4dca51a4-cb40-46ae-8545-d03f7d920305 service nova] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Received event network-changed-3ffa8001-d622-4cf0-865b-e5b16c4ddbcd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 693.340786] env[63355]: DEBUG nova.compute.manager [req-ae535077-6f45-42b6-9e2d-d43f8269f9ba req-4dca51a4-cb40-46ae-8545-d03f7d920305 service nova] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Refreshing instance network info cache due to event network-changed-3ffa8001-d622-4cf0-865b-e5b16c4ddbcd. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 693.340973] env[63355]: DEBUG oslo_concurrency.lockutils [req-ae535077-6f45-42b6-9e2d-d43f8269f9ba req-4dca51a4-cb40-46ae-8545-d03f7d920305 service nova] Acquiring lock "refresh_cache-087c0c34-2f44-4791-86e7-b1b3564f49d5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.554166] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.601150] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.707258] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.750444] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.806108] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.038427] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9bd587c-8a73-42b4-8316-7ebc691da605 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.046572] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899cdc9c-820f-431d-ab9d-025a19b6f737 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.076804] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fa3331-6f35-4945-8644-4b7adf4de0ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.084211] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac74b17-0a92-40c6-b8f3-af2a40a95ec2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.098307] env[63355]: DEBUG nova.compute.provider_tree [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.252140] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Releasing lock "refresh_cache-b275344f-695e-463b-8bfe-2bee97296349" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.252382] env[63355]: DEBUG nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 694.252559] env[63355]: DEBUG nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.252723] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.268034] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.307387] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Releasing lock "refresh_cache-087c0c34-2f44-4791-86e7-b1b3564f49d5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.308245] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.308443] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.308777] env[63355]: DEBUG oslo_concurrency.lockutils [req-ae535077-6f45-42b6-9e2d-d43f8269f9ba req-4dca51a4-cb40-46ae-8545-d03f7d920305 service nova] Acquired lock "refresh_cache-087c0c34-2f44-4791-86e7-b1b3564f49d5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.308988] env[63355]: DEBUG nova.network.neutron [req-ae535077-6f45-42b6-9e2d-d43f8269f9ba req-4dca51a4-cb40-46ae-8545-d03f7d920305 service nova] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Refreshing network info cache for port 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 694.310687] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66a5d41d-bc8c-4819-b799-76702d058574 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.319725] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc44884-fc91-4417-869d-091e5d8e069c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.344639] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 087c0c34-2f44-4791-86e7-b1b3564f49d5 could not be found. [ 694.344940] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.345182] env[63355]: INFO nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 694.345462] env[63355]: DEBUG oslo.service.loopingcall [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.345769] env[63355]: DEBUG nova.compute.manager [-] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.345899] env[63355]: DEBUG nova.network.neutron [-] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.363367] env[63355]: DEBUG nova.network.neutron [-] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.600874] env[63355]: DEBUG nova.scheduler.client.report [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.771120] env[63355]: DEBUG nova.network.neutron [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.832187] env[63355]: DEBUG nova.network.neutron [req-ae535077-6f45-42b6-9e2d-d43f8269f9ba req-4dca51a4-cb40-46ae-8545-d03f7d920305 service nova] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.869038] env[63355]: DEBUG nova.network.neutron [-] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.942736] env[63355]: DEBUG nova.network.neutron [req-ae535077-6f45-42b6-9e2d-d43f8269f9ba req-4dca51a4-cb40-46ae-8545-d03f7d920305 service nova] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.106365] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.088s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.106988] env[63355]: ERROR nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9d80c545-2d7e-43eb-9773-6ed1e023fa91, please check neutron logs for more information. [ 695.106988] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Traceback (most recent call last): [ 695.106988] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.106988] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] self.driver.spawn(context, instance, image_meta, [ 695.106988] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 695.106988] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.106988] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.106988] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] vm_ref = self.build_virtual_machine(instance, [ 695.106988] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.106988] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.106988] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] for vif in network_info: [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] return self._sync_wrapper(fn, *args, **kwargs) [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] self.wait() [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] self[:] = self._gt.wait() [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] return self._exit_event.wait() [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] current.throw(*self._exc) [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.107303] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] result = function(*args, **kwargs) [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] return func(*args, **kwargs) [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] raise e [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] nwinfo = self.network_api.allocate_for_instance( [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] created_port_ids = self._update_ports_for_instance( [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] with excutils.save_and_reraise_exception(): [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] self.force_reraise() [ 695.107615] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.107933] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] raise self.value [ 695.107933] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.107933] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] updated_port = self._update_port( [ 695.107933] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.107933] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] _ensure_no_port_binding_failure(port) [ 695.107933] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.107933] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] raise exception.PortBindingFailed(port_id=port['id']) [ 695.107933] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] nova.exception.PortBindingFailed: Binding failed for port 9d80c545-2d7e-43eb-9773-6ed1e023fa91, please check neutron logs for more information. [ 695.107933] env[63355]: ERROR nova.compute.manager [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] [ 695.107933] env[63355]: DEBUG nova.compute.utils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Binding failed for port 9d80c545-2d7e-43eb-9773-6ed1e023fa91, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 695.108969] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.567s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.115455] env[63355]: DEBUG nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Build of instance 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab was re-scheduled: Binding failed for port 9d80c545-2d7e-43eb-9773-6ed1e023fa91, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 695.117208] env[63355]: DEBUG nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 695.117459] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Acquiring lock "refresh_cache-3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.117607] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Acquired lock "refresh_cache-3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.117791] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.274391] env[63355]: INFO nova.compute.manager [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] [instance: b275344f-695e-463b-8bfe-2bee97296349] Took 1.02 seconds to deallocate network for instance. [ 695.372838] env[63355]: INFO nova.compute.manager [-] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Took 1.03 seconds to deallocate network for instance. [ 695.375422] env[63355]: DEBUG nova.compute.claims [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 695.375422] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.446703] env[63355]: DEBUG oslo_concurrency.lockutils [req-ae535077-6f45-42b6-9e2d-d43f8269f9ba req-4dca51a4-cb40-46ae-8545-d03f7d920305 service nova] Releasing lock "refresh_cache-087c0c34-2f44-4791-86e7-b1b3564f49d5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.447046] env[63355]: DEBUG nova.compute.manager [req-ae535077-6f45-42b6-9e2d-d43f8269f9ba req-4dca51a4-cb40-46ae-8545-d03f7d920305 service nova] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Received event network-vif-deleted-3ffa8001-d622-4cf0-865b-e5b16c4ddbcd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 695.648675] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.728397] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.091937] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d11d5f-2662-4ce7-b8c5-ae295d023593 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.099617] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30db0f46-1357-422c-84ec-b34e9d8c52ec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.130095] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e723081-7bb9-4292-9291-b30b334f1041 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.137804] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f58225-8020-49eb-8ce3-6008c9306245 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.150996] env[63355]: DEBUG nova.compute.provider_tree [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.235343] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Releasing lock "refresh_cache-3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.235586] env[63355]: DEBUG nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 696.235765] env[63355]: DEBUG nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.235957] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.267404] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.314022] env[63355]: INFO nova.scheduler.client.report [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Deleted allocations for instance b275344f-695e-463b-8bfe-2bee97296349 [ 696.365623] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "e49ba26c-486c-43aa-9595-231e7f31a522" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.365867] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "e49ba26c-486c-43aa-9595-231e7f31a522" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.653659] env[63355]: DEBUG nova.scheduler.client.report [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.771239] env[63355]: DEBUG nova.network.neutron [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.819788] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7fedf3a1-0993-4560-9579-8ef948fbb3d9 tempest-FloatingIPsAssociationTestJSON-757275121 tempest-FloatingIPsAssociationTestJSON-757275121-project-member] Lock "b275344f-695e-463b-8bfe-2bee97296349" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.914s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.159491] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.050s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.160142] env[63355]: ERROR nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e, please check neutron logs for more information. [ 697.160142] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Traceback (most recent call last): [ 697.160142] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.160142] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] self.driver.spawn(context, instance, image_meta, [ 697.160142] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 697.160142] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.160142] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.160142] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] vm_ref = self.build_virtual_machine(instance, [ 697.160142] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.160142] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.160142] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] for vif in network_info: [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] return self._sync_wrapper(fn, *args, **kwargs) [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] self.wait() [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] self[:] = self._gt.wait() [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] return self._exit_event.wait() [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] result = hub.switch() [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 697.160485] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] return self.greenlet.switch() [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] result = function(*args, **kwargs) [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] return func(*args, **kwargs) [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] raise e [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] nwinfo = self.network_api.allocate_for_instance( [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] created_port_ids = self._update_ports_for_instance( [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] with excutils.save_and_reraise_exception(): [ 697.160991] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] self.force_reraise() [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] raise self.value [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] updated_port = self._update_port( [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] _ensure_no_port_binding_failure(port) [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] raise exception.PortBindingFailed(port_id=port['id']) [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] nova.exception.PortBindingFailed: Binding failed for port 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e, please check neutron logs for more information. [ 697.162117] env[63355]: ERROR nova.compute.manager [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] [ 697.162500] env[63355]: DEBUG nova.compute.utils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Binding failed for port 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 697.162500] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.185s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.163559] env[63355]: INFO nova.compute.claims [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.166230] env[63355]: DEBUG nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Build of instance 0de6892a-5463-4d60-807c-0e3fca2accf3 was re-scheduled: Binding failed for port 71a2ce7f-a12c-4b69-ac20-e7d5b8a1126e, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 697.166655] env[63355]: DEBUG nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 697.166896] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-0de6892a-5463-4d60-807c-0e3fca2accf3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.167053] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-0de6892a-5463-4d60-807c-0e3fca2accf3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.167214] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.274070] env[63355]: INFO nova.compute.manager [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] [instance: 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab] Took 1.04 seconds to deallocate network for instance. [ 697.322885] env[63355]: DEBUG nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 697.689211] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.784485] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.851023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.289777] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-0de6892a-5463-4d60-807c-0e3fca2accf3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.289777] env[63355]: DEBUG nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 698.289777] env[63355]: DEBUG nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.290533] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.307168] env[63355]: INFO nova.scheduler.client.report [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Deleted allocations for instance 3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab [ 698.314989] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.733293] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0cea3a-f246-4991-9a40-2b6a46614457 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.741256] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38960bcf-8372-4542-a0e5-c2b3d4282196 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.774847] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a585ad0e-8b38-44eb-87cc-b03ef1c316a2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.783241] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b346d59-773a-48df-89a1-22aebc717bd4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.797575] env[63355]: DEBUG nova.compute.provider_tree [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.816669] env[63355]: DEBUG nova.network.neutron [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.822119] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5e255712-25f7-4c53-9ead-cab2bf414768 tempest-ServerRescueTestJSONUnderV235-2010528182 tempest-ServerRescueTestJSONUnderV235-2010528182-project-member] Lock "3dd3c8f5-f2ab-46b2-a6b8-beaeb906cbab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.474s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.301045] env[63355]: DEBUG nova.scheduler.client.report [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.319337] env[63355]: INFO nova.compute.manager [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 0de6892a-5463-4d60-807c-0e3fca2accf3] Took 1.03 seconds to deallocate network for instance. [ 699.325564] env[63355]: DEBUG nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 699.807211] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.645s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.807752] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 699.814284] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.151s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.861297] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.316959] env[63355]: DEBUG nova.compute.utils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 700.321436] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 700.321618] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 700.365723] env[63355]: INFO nova.scheduler.client.report [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted allocations for instance 0de6892a-5463-4d60-807c-0e3fca2accf3 [ 700.376366] env[63355]: DEBUG nova.policy [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b81f53d014ce4e5d937156d9a3e32c0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2bfdd9344794c6987e9915cf414b8d6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 700.825563] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 700.864649] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Successfully created port: 6b105f42-97df-4f8b-a719-3fcd47feeb49 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.881114] env[63355]: DEBUG oslo_concurrency.lockutils [None req-06a98934-5093-49e8-a209-5ff4b3d83c28 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "0de6892a-5463-4d60-807c-0e3fca2accf3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.298s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.938460] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900b0069-f2b4-4589-8c20-30eca2a16c4c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.951856] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde1edb0-6f42-47fd-9cfa-8403dda47eb0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.006921] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb3f35f-bb1e-4258-a730-89941c072966 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.017564] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d638b5-7f6d-4fdf-b586-b836b5f7d38b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.037992] env[63355]: DEBUG nova.compute.provider_tree [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.388115] env[63355]: DEBUG nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.542478] env[63355]: DEBUG nova.scheduler.client.report [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.844120] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 701.876508] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.876842] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.877263] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.877996] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.878612] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.878612] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.878719] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.878809] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.879040] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.879450] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.879450] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.880556] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af0a8b3-01b2-4815-afc0-d1da4cca0b3b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.896347] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed00b0a-6f90-4ce5-91aa-4b32bacec117 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.934702] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.047744] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.236s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.048884] env[63355]: ERROR nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f34f3878-f7a3-41e9-84ee-8e31f5577cda, please check neutron logs for more information. [ 702.048884] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Traceback (most recent call last): [ 702.048884] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.048884] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] self.driver.spawn(context, instance, image_meta, [ 702.048884] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 702.048884] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.048884] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.048884] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] vm_ref = self.build_virtual_machine(instance, [ 702.048884] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.048884] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.048884] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] for vif in network_info: [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] return self._sync_wrapper(fn, *args, **kwargs) [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] self.wait() [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] self[:] = self._gt.wait() [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] return self._exit_event.wait() [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] result = hub.switch() [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 702.049252] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] return self.greenlet.switch() [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] result = function(*args, **kwargs) [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] return func(*args, **kwargs) [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] raise e [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] nwinfo = self.network_api.allocate_for_instance( [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] created_port_ids = self._update_ports_for_instance( [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] with excutils.save_and_reraise_exception(): [ 702.049609] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] self.force_reraise() [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] raise self.value [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] updated_port = self._update_port( [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] _ensure_no_port_binding_failure(port) [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] raise exception.PortBindingFailed(port_id=port['id']) [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] nova.exception.PortBindingFailed: Binding failed for port f34f3878-f7a3-41e9-84ee-8e31f5577cda, please check neutron logs for more information. [ 702.049987] env[63355]: ERROR nova.compute.manager [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] [ 702.050324] env[63355]: DEBUG nova.compute.utils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Binding failed for port f34f3878-f7a3-41e9-84ee-8e31f5577cda, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 702.051459] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.671s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.059121] env[63355]: DEBUG nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Build of instance 2386393a-3006-4e6b-91e0-40b7ab8e3948 was re-scheduled: Binding failed for port f34f3878-f7a3-41e9-84ee-8e31f5577cda, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 702.059121] env[63355]: DEBUG nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 702.059121] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "refresh_cache-2386393a-3006-4e6b-91e0-40b7ab8e3948" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.059121] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "refresh_cache-2386393a-3006-4e6b-91e0-40b7ab8e3948" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.059121] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.392739] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "7dd28cb7-d249-4b8e-beb3-c959f15e912f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.393036] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "7dd28cb7-d249-4b8e-beb3-c959f15e912f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.537866] env[63355]: DEBUG nova.compute.manager [req-55a529bf-2074-430c-af9e-7856b32817e2 req-d6d3d4df-e3c5-403e-bcce-5785475820aa service nova] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Received event network-changed-6b105f42-97df-4f8b-a719-3fcd47feeb49 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.538121] env[63355]: DEBUG nova.compute.manager [req-55a529bf-2074-430c-af9e-7856b32817e2 req-d6d3d4df-e3c5-403e-bcce-5785475820aa service nova] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Refreshing instance network info cache due to event network-changed-6b105f42-97df-4f8b-a719-3fcd47feeb49. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 702.538649] env[63355]: DEBUG oslo_concurrency.lockutils [req-55a529bf-2074-430c-af9e-7856b32817e2 req-d6d3d4df-e3c5-403e-bcce-5785475820aa service nova] Acquiring lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.538912] env[63355]: DEBUG oslo_concurrency.lockutils [req-55a529bf-2074-430c-af9e-7856b32817e2 req-d6d3d4df-e3c5-403e-bcce-5785475820aa service nova] Acquired lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.539284] env[63355]: DEBUG nova.network.neutron [req-55a529bf-2074-430c-af9e-7856b32817e2 req-d6d3d4df-e3c5-403e-bcce-5785475820aa service nova] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Refreshing network info cache for port 6b105f42-97df-4f8b-a719-3fcd47feeb49 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 702.596092] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.753506] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.810150] env[63355]: ERROR nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6b105f42-97df-4f8b-a719-3fcd47feeb49, please check neutron logs for more information. [ 702.810150] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 702.810150] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.810150] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 702.810150] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.810150] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 702.810150] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.810150] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 702.810150] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.810150] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 702.810150] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.810150] env[63355]: ERROR nova.compute.manager raise self.value [ 702.810150] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.810150] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 702.810150] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.810150] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 702.811261] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.811261] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 702.811261] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6b105f42-97df-4f8b-a719-3fcd47feeb49, please check neutron logs for more information. [ 702.811261] env[63355]: ERROR nova.compute.manager [ 702.811261] env[63355]: Traceback (most recent call last): [ 702.811261] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 702.811261] env[63355]: listener.cb(fileno) [ 702.811261] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.811261] env[63355]: result = function(*args, **kwargs) [ 702.811261] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.811261] env[63355]: return func(*args, **kwargs) [ 702.811261] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.811261] env[63355]: raise e [ 702.811261] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.811261] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 702.811261] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.811261] env[63355]: created_port_ids = self._update_ports_for_instance( [ 702.811261] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.811261] env[63355]: with excutils.save_and_reraise_exception(): [ 702.811261] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.811261] env[63355]: self.force_reraise() [ 702.811261] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.811261] env[63355]: raise self.value [ 702.811261] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.811261] env[63355]: updated_port = self._update_port( [ 702.811261] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.811261] env[63355]: _ensure_no_port_binding_failure(port) [ 702.811261] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.811261] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 702.812247] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 6b105f42-97df-4f8b-a719-3fcd47feeb49, please check neutron logs for more information. [ 702.812247] env[63355]: Removing descriptor: 17 [ 702.812247] env[63355]: ERROR nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6b105f42-97df-4f8b-a719-3fcd47feeb49, please check neutron logs for more information. [ 702.812247] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Traceback (most recent call last): [ 702.812247] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 702.812247] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] yield resources [ 702.812247] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.812247] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] self.driver.spawn(context, instance, image_meta, [ 702.812247] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 702.812247] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.812247] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.812247] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] vm_ref = self.build_virtual_machine(instance, [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] for vif in network_info: [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] return self._sync_wrapper(fn, *args, **kwargs) [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] self.wait() [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] self[:] = self._gt.wait() [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] return self._exit_event.wait() [ 702.812606] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] result = hub.switch() [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] return self.greenlet.switch() [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] result = function(*args, **kwargs) [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] return func(*args, **kwargs) [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] raise e [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] nwinfo = self.network_api.allocate_for_instance( [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.812990] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] created_port_ids = self._update_ports_for_instance( [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] with excutils.save_and_reraise_exception(): [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] self.force_reraise() [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] raise self.value [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] updated_port = self._update_port( [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] _ensure_no_port_binding_failure(port) [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.813365] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] raise exception.PortBindingFailed(port_id=port['id']) [ 702.813878] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] nova.exception.PortBindingFailed: Binding failed for port 6b105f42-97df-4f8b-a719-3fcd47feeb49, please check neutron logs for more information. [ 702.813878] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] [ 702.813878] env[63355]: INFO nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Terminating instance [ 702.813878] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.076956] env[63355]: DEBUG nova.network.neutron [req-55a529bf-2074-430c-af9e-7856b32817e2 req-d6d3d4df-e3c5-403e-bcce-5785475820aa service nova] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.111545] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b2dcec-cdf3-4c31-a04f-d7633320deb0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.119197] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e3e91d-3c7e-4e6e-a70a-39dc5e9d0a08 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.150358] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afbe5c9-5b99-48f7-93fd-233131f6d13d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.158484] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1a8a2d-59f6-4fe8-99b4-79a8b5176114 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.171638] env[63355]: DEBUG nova.compute.provider_tree [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.256665] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "refresh_cache-2386393a-3006-4e6b-91e0-40b7ab8e3948" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.256901] env[63355]: DEBUG nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 703.257125] env[63355]: DEBUG nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.257298] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.259761] env[63355]: DEBUG nova.network.neutron [req-55a529bf-2074-430c-af9e-7856b32817e2 req-d6d3d4df-e3c5-403e-bcce-5785475820aa service nova] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.283835] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.674962] env[63355]: DEBUG nova.scheduler.client.report [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.762788] env[63355]: DEBUG oslo_concurrency.lockutils [req-55a529bf-2074-430c-af9e-7856b32817e2 req-d6d3d4df-e3c5-403e-bcce-5785475820aa service nova] Releasing lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.763291] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquired lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.763506] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.786021] env[63355]: DEBUG nova.network.neutron [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.180260] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.129s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.180963] env[63355]: ERROR nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1c6554f9-16ee-475e-a6cc-1c9e180c9ace, please check neutron logs for more information. [ 704.180963] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Traceback (most recent call last): [ 704.180963] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.180963] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] self.driver.spawn(context, instance, image_meta, [ 704.180963] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 704.180963] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.180963] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.180963] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] vm_ref = self.build_virtual_machine(instance, [ 704.180963] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.180963] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.180963] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] for vif in network_info: [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] return self._sync_wrapper(fn, *args, **kwargs) [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] self.wait() [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] self[:] = self._gt.wait() [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] return self._exit_event.wait() [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] result = hub.switch() [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 704.181324] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] return self.greenlet.switch() [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] result = function(*args, **kwargs) [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] return func(*args, **kwargs) [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] raise e [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] nwinfo = self.network_api.allocate_for_instance( [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] created_port_ids = self._update_ports_for_instance( [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] with excutils.save_and_reraise_exception(): [ 704.181645] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] self.force_reraise() [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] raise self.value [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] updated_port = self._update_port( [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] _ensure_no_port_binding_failure(port) [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] raise exception.PortBindingFailed(port_id=port['id']) [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] nova.exception.PortBindingFailed: Binding failed for port 1c6554f9-16ee-475e-a6cc-1c9e180c9ace, please check neutron logs for more information. [ 704.181990] env[63355]: ERROR nova.compute.manager [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] [ 704.182274] env[63355]: DEBUG nova.compute.utils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Binding failed for port 1c6554f9-16ee-475e-a6cc-1c9e180c9ace, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 704.182918] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.911s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.188171] env[63355]: DEBUG nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Build of instance 115e42ee-60cc-400d-8f4d-c23009f9091f was re-scheduled: Binding failed for port 1c6554f9-16ee-475e-a6cc-1c9e180c9ace, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 704.188171] env[63355]: DEBUG nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 704.188439] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquiring lock "refresh_cache-115e42ee-60cc-400d-8f4d-c23009f9091f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.188713] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Acquired lock "refresh_cache-115e42ee-60cc-400d-8f4d-c23009f9091f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.189832] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.291747] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.297188] env[63355]: INFO nova.compute.manager [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 2386393a-3006-4e6b-91e0-40b7ab8e3948] Took 1.04 seconds to deallocate network for instance. [ 704.427100] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.704605] env[63355]: DEBUG nova.compute.manager [req-4bd98b65-ecd4-4f7c-b1f1-310f5a725f11 req-dcfdfa25-1a47-4eea-9ee9-dbf5cea18e23 service nova] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Received event network-vif-deleted-6b105f42-97df-4f8b-a719-3fcd47feeb49 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.717754] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.760025] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "947a33d3-88c1-45ae-805c-eda2bc7626f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.760890] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "947a33d3-88c1-45ae-805c-eda2bc7626f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.808999] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.929701] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Releasing lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.930633] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 704.930633] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 704.930971] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52fa5661-157a-44c2-832b-8e664ddeac8e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.947916] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460bba96-3d29-4681-a587-fb9ffd26df5b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.976331] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 617b624f-b978-4f98-bbbc-befa006334e3 could not be found. [ 704.976589] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 704.976772] env[63355]: INFO nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 704.977031] env[63355]: DEBUG oslo.service.loopingcall [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.979539] env[63355]: DEBUG nova.compute.manager [-] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.979640] env[63355]: DEBUG nova.network.neutron [-] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.004458] env[63355]: DEBUG nova.network.neutron [-] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.265791] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c009b4b-c55b-4938-a140-33e66b939f86 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.276726] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f242a6b8-2297-49c5-8f05-c770527b59a2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.309377] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585ca18e-c1f4-4aa6-a099-781b3fd3141f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.316068] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Releasing lock "refresh_cache-115e42ee-60cc-400d-8f4d-c23009f9091f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.316068] env[63355]: DEBUG nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 705.316068] env[63355]: DEBUG nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.316068] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.325411] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf75df15-eff4-4aba-9316-bf817d634327 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.344325] env[63355]: DEBUG nova.compute.provider_tree [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.349583] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.356704] env[63355]: INFO nova.scheduler.client.report [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted allocations for instance 2386393a-3006-4e6b-91e0-40b7ab8e3948 [ 705.508682] env[63355]: DEBUG nova.network.neutron [-] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.813113] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquiring lock "3d918cab-6ed6-4a37-a024-28e3db1b779c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.813953] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lock "3d918cab-6ed6-4a37-a024-28e3db1b779c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.847822] env[63355]: DEBUG nova.scheduler.client.report [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.851081] env[63355]: DEBUG nova.network.neutron [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.866262] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c7c729e-0dcd-4e72-9dd0-d6bc666fcc0f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "2386393a-3006-4e6b-91e0-40b7ab8e3948" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.175s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.011885] env[63355]: INFO nova.compute.manager [-] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Took 1.03 seconds to deallocate network for instance. [ 706.014431] env[63355]: DEBUG nova.compute.claims [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 706.015125] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.353598] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.171s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.354298] env[63355]: ERROR nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c, please check neutron logs for more information. [ 706.354298] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] Traceback (most recent call last): [ 706.354298] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.354298] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] self.driver.spawn(context, instance, image_meta, [ 706.354298] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 706.354298] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.354298] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.354298] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] vm_ref = self.build_virtual_machine(instance, [ 706.354298] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.354298] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.354298] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] for vif in network_info: [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] return self._sync_wrapper(fn, *args, **kwargs) [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] self.wait() [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] self[:] = self._gt.wait() [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] return self._exit_event.wait() [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] result = hub.switch() [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 706.354933] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] return self.greenlet.switch() [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] result = function(*args, **kwargs) [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] return func(*args, **kwargs) [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] raise e [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] nwinfo = self.network_api.allocate_for_instance( [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] created_port_ids = self._update_ports_for_instance( [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] with excutils.save_and_reraise_exception(): [ 706.355660] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] self.force_reraise() [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] raise self.value [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] updated_port = self._update_port( [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] _ensure_no_port_binding_failure(port) [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] raise exception.PortBindingFailed(port_id=port['id']) [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] nova.exception.PortBindingFailed: Binding failed for port 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c, please check neutron logs for more information. [ 706.356294] env[63355]: ERROR nova.compute.manager [instance: b46098f3-d952-4a33-86fe-825fe4625008] [ 706.356766] env[63355]: DEBUG nova.compute.utils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Binding failed for port 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 706.358198] env[63355]: DEBUG nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Build of instance b46098f3-d952-4a33-86fe-825fe4625008 was re-scheduled: Binding failed for port 905828d8-0121-4c4e-83e0-cb7e6e4e6d2c, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 706.358642] env[63355]: DEBUG nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 706.358877] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Acquiring lock "refresh_cache-b46098f3-d952-4a33-86fe-825fe4625008" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.359038] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Acquired lock "refresh_cache-b46098f3-d952-4a33-86fe-825fe4625008" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.359204] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 706.362111] env[63355]: INFO nova.compute.manager [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] [instance: 115e42ee-60cc-400d-8f4d-c23009f9091f] Took 1.05 seconds to deallocate network for instance. [ 706.365082] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.725s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.368065] env[63355]: INFO nova.compute.claims [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.371742] env[63355]: DEBUG nova.compute.manager [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.893760] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.906241] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.012228] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.393411] env[63355]: INFO nova.scheduler.client.report [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Deleted allocations for instance 115e42ee-60cc-400d-8f4d-c23009f9091f [ 707.515512] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Releasing lock "refresh_cache-b46098f3-d952-4a33-86fe-825fe4625008" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.515736] env[63355]: DEBUG nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 707.515914] env[63355]: DEBUG nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.516099] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 707.540834] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.595584] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "5591bb55-83d7-4301-a3f9-fde945632344" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.595809] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "5591bb55-83d7-4301-a3f9-fde945632344" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.782575] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff7c64f-215f-4587-b0cd-f479c2299e69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.790033] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b25d03-c2dd-4f4d-9da3-f6b9841c4545 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.818642] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b55a11b2-b34e-4282-8fa1-2fafe11964e2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.825462] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443a5122-eb88-45d7-99ad-3bd96004804f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.838048] env[63355]: DEBUG nova.compute.provider_tree [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.904320] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4cee6489-42eb-4fe4-8d1d-11ab85a285e3 tempest-DeleteServersAdminTestJSON-1373063196 tempest-DeleteServersAdminTestJSON-1373063196-project-member] Lock "115e42ee-60cc-400d-8f4d-c23009f9091f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.030s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.043019] env[63355]: DEBUG nova.network.neutron [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.342521] env[63355]: DEBUG nova.scheduler.client.report [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.407195] env[63355]: DEBUG nova.compute.manager [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 708.546767] env[63355]: INFO nova.compute.manager [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] [instance: b46098f3-d952-4a33-86fe-825fe4625008] Took 1.03 seconds to deallocate network for instance. [ 708.849014] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.849014] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 708.855835] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.361s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.932149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.356766] env[63355]: DEBUG nova.compute.utils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.358131] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 709.358319] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 709.420407] env[63355]: DEBUG nova.policy [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b81f53d014ce4e5d937156d9a3e32c0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2bfdd9344794c6987e9915cf414b8d6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 709.580855] env[63355]: INFO nova.scheduler.client.report [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Deleted allocations for instance b46098f3-d952-4a33-86fe-825fe4625008 [ 709.821967] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c33e29-c651-4ef8-8e9d-9ac8cbc9653c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.829963] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5bb9332-3180-49e4-a9e6-3d9b27ea1464 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.863479] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.866598] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd5aa8d-ea39-45d0-8a4c-e619f516be03 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.874369] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2567670a-78cf-4790-bb1e-e30e8c982a84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.891555] env[63355]: DEBUG nova.compute.provider_tree [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.968019] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Successfully created port: e9eae179-ea5b-4a52-8f83-b85e299d41d3 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.090487] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6b0c823-d2a4-4373-af7d-06e7b27a4c74 tempest-ServerActionsTestJSON-2046750455 tempest-ServerActionsTestJSON-2046750455-project-member] Lock "b46098f3-d952-4a33-86fe-825fe4625008" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.015s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.394572] env[63355]: DEBUG nova.scheduler.client.report [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.594152] env[63355]: DEBUG nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 710.879829] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 710.902634] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.047s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.903589] env[63355]: ERROR nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b75086db-4824-445c-8b78-912828bec10e, please check neutron logs for more information. [ 710.903589] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Traceback (most recent call last): [ 710.903589] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.903589] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] self.driver.spawn(context, instance, image_meta, [ 710.903589] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 710.903589] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.903589] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.903589] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] vm_ref = self.build_virtual_machine(instance, [ 710.903589] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.903589] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.903589] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] for vif in network_info: [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] return self._sync_wrapper(fn, *args, **kwargs) [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] self.wait() [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] self[:] = self._gt.wait() [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] return self._exit_event.wait() [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] result = hub.switch() [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 710.903925] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] return self.greenlet.switch() [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] result = function(*args, **kwargs) [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] return func(*args, **kwargs) [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] raise e [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] nwinfo = self.network_api.allocate_for_instance( [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] created_port_ids = self._update_ports_for_instance( [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] with excutils.save_and_reraise_exception(): [ 710.904285] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] self.force_reraise() [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] raise self.value [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] updated_port = self._update_port( [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] _ensure_no_port_binding_failure(port) [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] raise exception.PortBindingFailed(port_id=port['id']) [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] nova.exception.PortBindingFailed: Binding failed for port b75086db-4824-445c-8b78-912828bec10e, please check neutron logs for more information. [ 710.904619] env[63355]: ERROR nova.compute.manager [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] [ 710.905141] env[63355]: DEBUG nova.compute.utils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Binding failed for port b75086db-4824-445c-8b78-912828bec10e, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 710.907093] env[63355]: DEBUG nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Build of instance 732ff6d3-2de9-4e6b-887d-5329ba83443c was re-scheduled: Binding failed for port b75086db-4824-445c-8b78-912828bec10e, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 710.907640] env[63355]: DEBUG nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 710.907955] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-732ff6d3-2de9-4e6b-887d-5329ba83443c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.908194] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-732ff6d3-2de9-4e6b-887d-5329ba83443c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.908562] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.912983] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.235s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.916205] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.916205] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 710.916205] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.313s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.916205] env[63355]: DEBUG nova.objects.instance [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lazy-loading 'resources' on Instance uuid def90353-ae90-4e1e-9fe1-8e5459a7309d {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 710.916205] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59faa229-60f0-4365-a863-ebf527f7bc3b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.923722] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.923980] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.924167] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.924404] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.924617] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.924760] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.924948] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.925864] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.925864] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.925864] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.925864] env[63355]: DEBUG nova.virt.hardware [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.927637] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6dc684-22a4-4e25-9894-d5a562efd6ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.941141] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98dfacd-08a8-41c8-93ac-c22e827f79db {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.946836] env[63355]: DEBUG nova.compute.manager [req-289ab38e-31be-4622-bd6e-4febd6c7d718 req-c088a570-5aa5-4c9b-8eb0-c57917400ea3 service nova] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Received event network-changed-e9eae179-ea5b-4a52-8f83-b85e299d41d3 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.947110] env[63355]: DEBUG nova.compute.manager [req-289ab38e-31be-4622-bd6e-4febd6c7d718 req-c088a570-5aa5-4c9b-8eb0-c57917400ea3 service nova] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Refreshing instance network info cache due to event network-changed-e9eae179-ea5b-4a52-8f83-b85e299d41d3. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 710.947365] env[63355]: DEBUG oslo_concurrency.lockutils [req-289ab38e-31be-4622-bd6e-4febd6c7d718 req-c088a570-5aa5-4c9b-8eb0-c57917400ea3 service nova] Acquiring lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.947562] env[63355]: DEBUG oslo_concurrency.lockutils [req-289ab38e-31be-4622-bd6e-4febd6c7d718 req-c088a570-5aa5-4c9b-8eb0-c57917400ea3 service nova] Acquired lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.947804] env[63355]: DEBUG nova.network.neutron [req-289ab38e-31be-4622-bd6e-4febd6c7d718 req-c088a570-5aa5-4c9b-8eb0-c57917400ea3 service nova] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Refreshing network info cache for port e9eae179-ea5b-4a52-8f83-b85e299d41d3 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 710.951953] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12958704-f56d-4bf1-aee3-3bcc7d6bf6e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.977940] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbc5232-dd55-4357-b75f-eb8a5f16845e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.984692] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5dcf18b-6baf-44d9-b427-2ae328bdde26 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.014462] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181290MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 711.014625] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.120428] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.352452] env[63355]: ERROR nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e9eae179-ea5b-4a52-8f83-b85e299d41d3, please check neutron logs for more information. [ 711.352452] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 711.352452] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.352452] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 711.352452] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.352452] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 711.352452] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.352452] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 711.352452] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.352452] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 711.352452] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.352452] env[63355]: ERROR nova.compute.manager raise self.value [ 711.352452] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.352452] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 711.352452] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.352452] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 711.352887] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.352887] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 711.352887] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e9eae179-ea5b-4a52-8f83-b85e299d41d3, please check neutron logs for more information. [ 711.352887] env[63355]: ERROR nova.compute.manager [ 711.352887] env[63355]: Traceback (most recent call last): [ 711.352887] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 711.352887] env[63355]: listener.cb(fileno) [ 711.352887] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.352887] env[63355]: result = function(*args, **kwargs) [ 711.352887] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.352887] env[63355]: return func(*args, **kwargs) [ 711.352887] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.352887] env[63355]: raise e [ 711.352887] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.352887] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 711.352887] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.352887] env[63355]: created_port_ids = self._update_ports_for_instance( [ 711.352887] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.352887] env[63355]: with excutils.save_and_reraise_exception(): [ 711.352887] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.352887] env[63355]: self.force_reraise() [ 711.352887] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.352887] env[63355]: raise self.value [ 711.352887] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.352887] env[63355]: updated_port = self._update_port( [ 711.352887] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.352887] env[63355]: _ensure_no_port_binding_failure(port) [ 711.352887] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.352887] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 711.353612] env[63355]: nova.exception.PortBindingFailed: Binding failed for port e9eae179-ea5b-4a52-8f83-b85e299d41d3, please check neutron logs for more information. [ 711.353612] env[63355]: Removing descriptor: 18 [ 711.353612] env[63355]: ERROR nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e9eae179-ea5b-4a52-8f83-b85e299d41d3, please check neutron logs for more information. [ 711.353612] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Traceback (most recent call last): [ 711.353612] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 711.353612] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] yield resources [ 711.353612] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.353612] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] self.driver.spawn(context, instance, image_meta, [ 711.353612] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 711.353612] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.353612] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.353612] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] vm_ref = self.build_virtual_machine(instance, [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] for vif in network_info: [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] return self._sync_wrapper(fn, *args, **kwargs) [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] self.wait() [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] self[:] = self._gt.wait() [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] return self._exit_event.wait() [ 711.353954] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] result = hub.switch() [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] return self.greenlet.switch() [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] result = function(*args, **kwargs) [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] return func(*args, **kwargs) [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] raise e [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] nwinfo = self.network_api.allocate_for_instance( [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.354317] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] created_port_ids = self._update_ports_for_instance( [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] with excutils.save_and_reraise_exception(): [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] self.force_reraise() [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] raise self.value [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] updated_port = self._update_port( [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] _ensure_no_port_binding_failure(port) [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.354709] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] raise exception.PortBindingFailed(port_id=port['id']) [ 711.355095] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] nova.exception.PortBindingFailed: Binding failed for port e9eae179-ea5b-4a52-8f83-b85e299d41d3, please check neutron logs for more information. [ 711.355095] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] [ 711.355095] env[63355]: INFO nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Terminating instance [ 711.356275] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.454162] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.473781] env[63355]: DEBUG nova.network.neutron [req-289ab38e-31be-4622-bd6e-4febd6c7d718 req-c088a570-5aa5-4c9b-8eb0-c57917400ea3 service nova] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.596556] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.612465] env[63355]: DEBUG nova.network.neutron [req-289ab38e-31be-4622-bd6e-4febd6c7d718 req-c088a570-5aa5-4c9b-8eb0-c57917400ea3 service nova] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.900220] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa02acd1-b08c-4dfc-9370-a6c2d7a8b8a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.909044] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd4ee7f-47f7-49d1-83e0-d4a3fb6e6672 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.941099] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee6ce42-c248-43af-843d-a5e1464c12a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.948851] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14562e8c-5e35-4ebe-acea-c9f3ba691fb8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.963070] env[63355]: DEBUG nova.compute.provider_tree [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.102619] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-732ff6d3-2de9-4e6b-887d-5329ba83443c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.103585] env[63355]: DEBUG nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 712.103585] env[63355]: DEBUG nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.103585] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.115113] env[63355]: DEBUG oslo_concurrency.lockutils [req-289ab38e-31be-4622-bd6e-4febd6c7d718 req-c088a570-5aa5-4c9b-8eb0-c57917400ea3 service nova] Releasing lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.115491] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquired lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.117133] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.125148] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.466607] env[63355]: DEBUG nova.scheduler.client.report [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.630011] env[63355]: DEBUG nova.network.neutron [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.633984] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.764910] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.973364] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.059s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.975816] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.600s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.002485] env[63355]: INFO nova.scheduler.client.report [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Deleted allocations for instance def90353-ae90-4e1e-9fe1-8e5459a7309d [ 713.070602] env[63355]: DEBUG nova.compute.manager [req-4d8e3698-4201-4365-b4aa-43569a23460f req-102a3ba8-563a-4c46-8818-706975a310a3 service nova] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Received event network-vif-deleted-e9eae179-ea5b-4a52-8f83-b85e299d41d3 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.131437] env[63355]: INFO nova.compute.manager [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 732ff6d3-2de9-4e6b-887d-5329ba83443c] Took 1.03 seconds to deallocate network for instance. [ 713.267869] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Releasing lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.268617] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 713.268617] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 713.268814] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dad69c81-6f52-4f0f-9c83-c01b5cbc8167 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.280126] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4737ae0-34b8-488d-b021-931a6b269cbe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.308054] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 12f9b2b4-3c78-43c2-bb9b-d44b81c83123 could not be found. [ 713.308054] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.308054] env[63355]: INFO nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Took 0.04 seconds to destroy the instance on the hypervisor. [ 713.308054] env[63355]: DEBUG oslo.service.loopingcall [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.308054] env[63355]: DEBUG nova.compute.manager [-] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.308054] env[63355]: DEBUG nova.network.neutron [-] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.325241] env[63355]: DEBUG nova.network.neutron [-] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.516742] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f58b571f-1b72-4349-a1da-01143f5ae9a1 tempest-ServerShowV247Test-296090945 tempest-ServerShowV247Test-296090945-project-member] Lock "def90353-ae90-4e1e-9fe1-8e5459a7309d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.858s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.827254] env[63355]: DEBUG nova.network.neutron [-] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.969098] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd3fee9-5d3e-44cf-98da-8609710dc9eb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.979160] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8774bb6-966d-4acb-85a2-1aab7d062358 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.018116] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993e48d5-ec46-42e2-abe4-9a32cd9bf016 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.026150] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f028f50b-55b9-44b3-8917-3794e07e89a4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.038258] env[63355]: DEBUG nova.compute.provider_tree [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.161850] env[63355]: INFO nova.scheduler.client.report [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Deleted allocations for instance 732ff6d3-2de9-4e6b-887d-5329ba83443c [ 714.330076] env[63355]: INFO nova.compute.manager [-] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Took 1.02 seconds to deallocate network for instance. [ 714.332478] env[63355]: DEBUG nova.compute.claims [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 714.332663] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.541732] env[63355]: DEBUG nova.scheduler.client.report [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.670449] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a90b2ff7-eea3-458a-a08d-aafffbbff6b2 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "732ff6d3-2de9-4e6b-887d-5329ba83443c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.248s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.047318] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.072s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.048014] env[63355]: ERROR nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd, please check neutron logs for more information. [ 715.048014] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Traceback (most recent call last): [ 715.048014] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.048014] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] self.driver.spawn(context, instance, image_meta, [ 715.048014] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 715.048014] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.048014] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.048014] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] vm_ref = self.build_virtual_machine(instance, [ 715.048014] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.048014] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.048014] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] for vif in network_info: [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] return self._sync_wrapper(fn, *args, **kwargs) [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] self.wait() [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] self[:] = self._gt.wait() [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] return self._exit_event.wait() [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] result = hub.switch() [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 715.048303] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] return self.greenlet.switch() [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] result = function(*args, **kwargs) [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] return func(*args, **kwargs) [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] raise e [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] nwinfo = self.network_api.allocate_for_instance( [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] created_port_ids = self._update_ports_for_instance( [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] with excutils.save_and_reraise_exception(): [ 715.048677] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] self.force_reraise() [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] raise self.value [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] updated_port = self._update_port( [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] _ensure_no_port_binding_failure(port) [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] raise exception.PortBindingFailed(port_id=port['id']) [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] nova.exception.PortBindingFailed: Binding failed for port 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd, please check neutron logs for more information. [ 715.049061] env[63355]: ERROR nova.compute.manager [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] [ 715.049371] env[63355]: DEBUG nova.compute.utils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Binding failed for port 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 715.049970] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.199s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.051475] env[63355]: INFO nova.compute.claims [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.054963] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Build of instance 087c0c34-2f44-4791-86e7-b1b3564f49d5 was re-scheduled: Binding failed for port 3ffa8001-d622-4cf0-865b-e5b16c4ddbcd, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 715.055898] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 715.055898] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "refresh_cache-087c0c34-2f44-4791-86e7-b1b3564f49d5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.055898] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquired lock "refresh_cache-087c0c34-2f44-4791-86e7-b1b3564f49d5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.056068] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.176971] env[63355]: DEBUG nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.577588] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.690315] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.713718] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.193175] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Releasing lock "refresh_cache-087c0c34-2f44-4791-86e7-b1b3564f49d5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.193580] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 716.193580] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.193742] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.213414] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.453869] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5924bf9f-7983-4f92-9190-149dc2ed556b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.462070] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142ad641-0005-413f-9372-7e26d7715c44 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.492486] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e390b94-4d3a-4c06-bf00-4652e64b9cf1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.500032] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f9ae4a-d4dd-417c-9ebf-68771d4c7ab2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.513356] env[63355]: DEBUG nova.compute.provider_tree [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.716584] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.905646] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "386e847e-967b-4247-9730-cdc5ac251474" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.905917] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "386e847e-967b-4247-9730-cdc5ac251474" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.016553] env[63355]: DEBUG nova.scheduler.client.report [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.220012] env[63355]: INFO nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 087c0c34-2f44-4791-86e7-b1b3564f49d5] Took 1.03 seconds to deallocate network for instance. [ 717.522464] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.522464] env[63355]: DEBUG nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 717.525552] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.665s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.526908] env[63355]: INFO nova.compute.claims [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.027258] env[63355]: DEBUG nova.compute.utils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.028646] env[63355]: DEBUG nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 718.028814] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 718.086988] env[63355]: DEBUG nova.policy [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dde65519973e474cb6d4acb7d37ed41a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f28f90c76b8345c5991b6b8bd54f0237', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 718.258157] env[63355]: INFO nova.scheduler.client.report [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Deleted allocations for instance 087c0c34-2f44-4791-86e7-b1b3564f49d5 [ 718.366763] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Successfully created port: 47293971-a17c-4da3-a20d-458ea560a914 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 718.532392] env[63355]: DEBUG nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 718.767874] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "087c0c34-2f44-4791-86e7-b1b3564f49d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.705s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.982904] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36be59f0-7018-4a49-b210-ad7fe40e7198 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.991387] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251784c9-66fa-49a2-ae33-7168f131b995 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.020897] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc819955-d638-4166-bca0-e9e2d67dc6be {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.028680] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6cd28a-3598-429e-a494-61ebac97253a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.047434] env[63355]: DEBUG nova.compute.provider_tree [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.137505] env[63355]: DEBUG nova.compute.manager [req-9b1677e7-ea44-4851-bb99-23db1363a6a2 req-103735e2-1aec-4515-8148-6f2bb45f8a2e service nova] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Received event network-changed-47293971-a17c-4da3-a20d-458ea560a914 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.137729] env[63355]: DEBUG nova.compute.manager [req-9b1677e7-ea44-4851-bb99-23db1363a6a2 req-103735e2-1aec-4515-8148-6f2bb45f8a2e service nova] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Refreshing instance network info cache due to event network-changed-47293971-a17c-4da3-a20d-458ea560a914. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 719.138018] env[63355]: DEBUG oslo_concurrency.lockutils [req-9b1677e7-ea44-4851-bb99-23db1363a6a2 req-103735e2-1aec-4515-8148-6f2bb45f8a2e service nova] Acquiring lock "refresh_cache-ad05eda3-efd7-4040-be9c-6a47df02889b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.138192] env[63355]: DEBUG oslo_concurrency.lockutils [req-9b1677e7-ea44-4851-bb99-23db1363a6a2 req-103735e2-1aec-4515-8148-6f2bb45f8a2e service nova] Acquired lock "refresh_cache-ad05eda3-efd7-4040-be9c-6a47df02889b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.138437] env[63355]: DEBUG nova.network.neutron [req-9b1677e7-ea44-4851-bb99-23db1363a6a2 req-103735e2-1aec-4515-8148-6f2bb45f8a2e service nova] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Refreshing network info cache for port 47293971-a17c-4da3-a20d-458ea560a914 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 719.268409] env[63355]: ERROR nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 47293971-a17c-4da3-a20d-458ea560a914, please check neutron logs for more information. [ 719.268409] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 719.268409] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.268409] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 719.268409] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.268409] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 719.268409] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.268409] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 719.268409] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.268409] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 719.268409] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.268409] env[63355]: ERROR nova.compute.manager raise self.value [ 719.268409] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.268409] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 719.268409] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.268409] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 719.269521] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.269521] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 719.269521] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 47293971-a17c-4da3-a20d-458ea560a914, please check neutron logs for more information. [ 719.269521] env[63355]: ERROR nova.compute.manager [ 719.269521] env[63355]: Traceback (most recent call last): [ 719.269521] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 719.269521] env[63355]: listener.cb(fileno) [ 719.269521] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.269521] env[63355]: result = function(*args, **kwargs) [ 719.269521] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.269521] env[63355]: return func(*args, **kwargs) [ 719.269521] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.269521] env[63355]: raise e [ 719.269521] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.269521] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 719.269521] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.269521] env[63355]: created_port_ids = self._update_ports_for_instance( [ 719.269521] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.269521] env[63355]: with excutils.save_and_reraise_exception(): [ 719.269521] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.269521] env[63355]: self.force_reraise() [ 719.269521] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.269521] env[63355]: raise self.value [ 719.269521] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.269521] env[63355]: updated_port = self._update_port( [ 719.269521] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.269521] env[63355]: _ensure_no_port_binding_failure(port) [ 719.269521] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.269521] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 719.272275] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 47293971-a17c-4da3-a20d-458ea560a914, please check neutron logs for more information. [ 719.272275] env[63355]: Removing descriptor: 17 [ 719.272275] env[63355]: DEBUG nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.551021] env[63355]: DEBUG nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 719.554711] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "12f9b2b4-3c78-43c2-bb9b-d44b81c83123" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.555400] env[63355]: DEBUG nova.scheduler.client.report [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.572082] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:25:21Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2d9c0613-6806-4206-aa39-e98bfe7d5439',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-18272180',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 719.572346] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 719.572501] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 719.573109] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 719.573109] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 719.573109] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 719.573245] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 719.573320] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 719.573475] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 719.573637] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 719.573807] env[63355]: DEBUG nova.virt.hardware [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 719.574701] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd492d7b-3ff3-4362-a42b-011f9796ea1f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.583761] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3c36d7-0764-4a0a-ae9f-3f9e7f291799 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.597989] env[63355]: ERROR nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 47293971-a17c-4da3-a20d-458ea560a914, please check neutron logs for more information. [ 719.597989] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Traceback (most recent call last): [ 719.597989] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 719.597989] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] yield resources [ 719.597989] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.597989] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] self.driver.spawn(context, instance, image_meta, [ 719.597989] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 719.597989] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.597989] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.597989] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] vm_ref = self.build_virtual_machine(instance, [ 719.597989] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] for vif in network_info: [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] return self._sync_wrapper(fn, *args, **kwargs) [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] self.wait() [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] self[:] = self._gt.wait() [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] return self._exit_event.wait() [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 719.598393] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] current.throw(*self._exc) [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] result = function(*args, **kwargs) [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] return func(*args, **kwargs) [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] raise e [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] nwinfo = self.network_api.allocate_for_instance( [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] created_port_ids = self._update_ports_for_instance( [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] with excutils.save_and_reraise_exception(): [ 719.598814] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] self.force_reraise() [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] raise self.value [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] updated_port = self._update_port( [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] _ensure_no_port_binding_failure(port) [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] raise exception.PortBindingFailed(port_id=port['id']) [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] nova.exception.PortBindingFailed: Binding failed for port 47293971-a17c-4da3-a20d-458ea560a914, please check neutron logs for more information. [ 719.599338] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] [ 719.599338] env[63355]: INFO nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Terminating instance [ 719.600517] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquiring lock "refresh_cache-ad05eda3-efd7-4040-be9c-6a47df02889b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.656797] env[63355]: DEBUG nova.network.neutron [req-9b1677e7-ea44-4851-bb99-23db1363a6a2 req-103735e2-1aec-4515-8148-6f2bb45f8a2e service nova] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.704531] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "617b624f-b978-4f98-bbbc-befa006334e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.710636] env[63355]: DEBUG nova.network.neutron [req-9b1677e7-ea44-4851-bb99-23db1363a6a2 req-103735e2-1aec-4515-8148-6f2bb45f8a2e service nova] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.791810] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.061057] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.062030] env[63355]: DEBUG nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 720.064498] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.130s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.065948] env[63355]: INFO nova.compute.claims [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.214028] env[63355]: DEBUG oslo_concurrency.lockutils [req-9b1677e7-ea44-4851-bb99-23db1363a6a2 req-103735e2-1aec-4515-8148-6f2bb45f8a2e service nova] Releasing lock "refresh_cache-ad05eda3-efd7-4040-be9c-6a47df02889b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.214852] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquired lock "refresh_cache-ad05eda3-efd7-4040-be9c-6a47df02889b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.214852] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.516579] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "0b660647-697a-445b-bca4-82ef6975ea75" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.516833] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "0b660647-697a-445b-bca4-82ef6975ea75" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.571010] env[63355]: DEBUG nova.compute.utils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.575024] env[63355]: DEBUG nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 720.575024] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 720.626764] env[63355]: DEBUG nova.policy [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '353154a4353f491d9e5d4f184690486a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ff6e81b016d44d79afafaab625cbe7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 720.733361] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.765402] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.765637] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.803801] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.951650] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Successfully created port: 6973d5f6-e86d-4734-85ff-ff458f776431 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.075457] env[63355]: DEBUG nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 721.174432] env[63355]: DEBUG nova.compute.manager [req-117ad17a-2433-4cda-b115-4255e989552d req-e83ace0c-3409-42ed-b459-3fc12c8900db service nova] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Received event network-vif-deleted-47293971-a17c-4da3-a20d-458ea560a914 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.307235] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Releasing lock "refresh_cache-ad05eda3-efd7-4040-be9c-6a47df02889b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.307675] env[63355]: DEBUG nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 721.307881] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 721.310844] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9bc08336-3eb4-4187-959b-c86b12b83f83 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.322265] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34337f64-a833-498c-82b9-60a63f9724e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.348579] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad05eda3-efd7-4040-be9c-6a47df02889b could not be found. [ 721.348579] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 721.348579] env[63355]: INFO nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 721.348579] env[63355]: DEBUG oslo.service.loopingcall [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.350801] env[63355]: DEBUG nova.compute.manager [-] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 721.350982] env[63355]: DEBUG nova.network.neutron [-] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.386330] env[63355]: DEBUG nova.network.neutron [-] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.546067] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27219e41-c1ed-42dd-b6e2-cc894e0c8469 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.555542] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d34c53-1e0c-4367-928d-0415eb2b42f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.589307] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832a0933-2180-4186-ac67-278f25ae8fa1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.597050] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a8892a-9036-43e6-aad8-a53618860eba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.614019] env[63355]: DEBUG nova.compute.provider_tree [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.806367] env[63355]: ERROR nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6973d5f6-e86d-4734-85ff-ff458f776431, please check neutron logs for more information. [ 721.806367] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 721.806367] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.806367] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 721.806367] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.806367] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 721.806367] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.806367] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 721.806367] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.806367] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 721.806367] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.806367] env[63355]: ERROR nova.compute.manager raise self.value [ 721.806367] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.806367] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 721.806367] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.806367] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 721.807233] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.807233] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 721.807233] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6973d5f6-e86d-4734-85ff-ff458f776431, please check neutron logs for more information. [ 721.807233] env[63355]: ERROR nova.compute.manager [ 721.807233] env[63355]: Traceback (most recent call last): [ 721.807233] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 721.807233] env[63355]: listener.cb(fileno) [ 721.807233] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.807233] env[63355]: result = function(*args, **kwargs) [ 721.807233] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.807233] env[63355]: return func(*args, **kwargs) [ 721.807233] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.807233] env[63355]: raise e [ 721.807233] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.807233] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 721.807233] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.807233] env[63355]: created_port_ids = self._update_ports_for_instance( [ 721.807233] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.807233] env[63355]: with excutils.save_and_reraise_exception(): [ 721.807233] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.807233] env[63355]: self.force_reraise() [ 721.807233] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.807233] env[63355]: raise self.value [ 721.807233] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.807233] env[63355]: updated_port = self._update_port( [ 721.807233] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.807233] env[63355]: _ensure_no_port_binding_failure(port) [ 721.807233] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.807233] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 721.808740] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 6973d5f6-e86d-4734-85ff-ff458f776431, please check neutron logs for more information. [ 721.808740] env[63355]: Removing descriptor: 17 [ 721.894127] env[63355]: DEBUG nova.network.neutron [-] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.094045] env[63355]: DEBUG nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 722.117624] env[63355]: DEBUG nova.scheduler.client.report [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.123054] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 722.123294] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 722.123450] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.123626] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 722.123773] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.123917] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 722.124212] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 722.124378] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 722.124541] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 722.124700] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 722.124868] env[63355]: DEBUG nova.virt.hardware [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.125745] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1163acfd-2575-4801-acab-5741aec9672e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.134645] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7adfa729-cfb7-4806-b4f4-30dd95f1d219 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.149095] env[63355]: ERROR nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6973d5f6-e86d-4734-85ff-ff458f776431, please check neutron logs for more information. [ 722.149095] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Traceback (most recent call last): [ 722.149095] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 722.149095] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] yield resources [ 722.149095] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.149095] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] self.driver.spawn(context, instance, image_meta, [ 722.149095] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 722.149095] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.149095] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.149095] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] vm_ref = self.build_virtual_machine(instance, [ 722.149095] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] for vif in network_info: [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] return self._sync_wrapper(fn, *args, **kwargs) [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] self.wait() [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] self[:] = self._gt.wait() [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] return self._exit_event.wait() [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.149493] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] current.throw(*self._exc) [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] result = function(*args, **kwargs) [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] return func(*args, **kwargs) [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] raise e [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] nwinfo = self.network_api.allocate_for_instance( [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] created_port_ids = self._update_ports_for_instance( [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] with excutils.save_and_reraise_exception(): [ 722.149879] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] self.force_reraise() [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] raise self.value [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] updated_port = self._update_port( [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] _ensure_no_port_binding_failure(port) [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] raise exception.PortBindingFailed(port_id=port['id']) [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] nova.exception.PortBindingFailed: Binding failed for port 6973d5f6-e86d-4734-85ff-ff458f776431, please check neutron logs for more information. [ 722.150267] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] [ 722.150267] env[63355]: INFO nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Terminating instance [ 722.151343] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Acquiring lock "refresh_cache-9478ce7a-99c6-4e95-82b0-f8a71ce94a90" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.151503] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Acquired lock "refresh_cache-9478ce7a-99c6-4e95-82b0-f8a71ce94a90" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.151665] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 722.396719] env[63355]: INFO nova.compute.manager [-] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Took 1.05 seconds to deallocate network for instance. [ 722.398364] env[63355]: DEBUG nova.compute.claims [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 722.398544] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.629702] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.565s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.630329] env[63355]: DEBUG nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.632785] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.618s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.670518] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.720242] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.137376] env[63355]: DEBUG nova.compute.utils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.142009] env[63355]: DEBUG nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 723.142190] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 723.194057] env[63355]: DEBUG nova.policy [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7123da3ee74b497395f1681e460aaaca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b6963b8e16b4986a4545914b75a38ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.222114] env[63355]: DEBUG nova.compute.manager [req-1b0a08b2-7fcd-498f-8412-563b16e6c986 req-8d70cdb6-1137-4bc8-a75f-83bce0c681ed service nova] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Received event network-changed-6973d5f6-e86d-4734-85ff-ff458f776431 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.222325] env[63355]: DEBUG nova.compute.manager [req-1b0a08b2-7fcd-498f-8412-563b16e6c986 req-8d70cdb6-1137-4bc8-a75f-83bce0c681ed service nova] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Refreshing instance network info cache due to event network-changed-6973d5f6-e86d-4734-85ff-ff458f776431. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 723.222513] env[63355]: DEBUG oslo_concurrency.lockutils [req-1b0a08b2-7fcd-498f-8412-563b16e6c986 req-8d70cdb6-1137-4bc8-a75f-83bce0c681ed service nova] Acquiring lock "refresh_cache-9478ce7a-99c6-4e95-82b0-f8a71ce94a90" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.222888] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Releasing lock "refresh_cache-9478ce7a-99c6-4e95-82b0-f8a71ce94a90" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.223271] env[63355]: DEBUG nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 723.223456] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 723.223995] env[63355]: DEBUG oslo_concurrency.lockutils [req-1b0a08b2-7fcd-498f-8412-563b16e6c986 req-8d70cdb6-1137-4bc8-a75f-83bce0c681ed service nova] Acquired lock "refresh_cache-9478ce7a-99c6-4e95-82b0-f8a71ce94a90" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.224356] env[63355]: DEBUG nova.network.neutron [req-1b0a08b2-7fcd-498f-8412-563b16e6c986 req-8d70cdb6-1137-4bc8-a75f-83bce0c681ed service nova] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Refreshing network info cache for port 6973d5f6-e86d-4734-85ff-ff458f776431 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 723.226182] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad24f93c-3294-47fe-bb3e-84a9f615c51a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.237536] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe105b7-b95a-40ec-af7d-cf71af8c4491 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.263518] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9478ce7a-99c6-4e95-82b0-f8a71ce94a90 could not be found. [ 723.263753] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 723.263929] env[63355]: INFO nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Took 0.04 seconds to destroy the instance on the hypervisor. [ 723.264246] env[63355]: DEBUG oslo.service.loopingcall [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 723.266747] env[63355]: DEBUG nova.compute.manager [-] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.266849] env[63355]: DEBUG nova.network.neutron [-] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 723.286667] env[63355]: DEBUG nova.network.neutron [-] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.469669] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Successfully created port: ac49af64-6ba9-4980-b8dc-ac660a40adae {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.583442] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7280098-1457-44f6-9cec-b6d26770cc9e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.590963] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f9c8a7-dd8e-41e6-8ebc-70f9d647ba12 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.620334] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6cfb27a-dd78-4d39-a265-69946eb68908 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.627590] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce105aab-dde2-4f00-a402-fbcb1321da27 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.640615] env[63355]: DEBUG nova.compute.provider_tree [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.644453] env[63355]: DEBUG nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.754842] env[63355]: DEBUG nova.network.neutron [req-1b0a08b2-7fcd-498f-8412-563b16e6c986 req-8d70cdb6-1137-4bc8-a75f-83bce0c681ed service nova] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.788527] env[63355]: DEBUG nova.network.neutron [-] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.811185] env[63355]: DEBUG nova.network.neutron [req-1b0a08b2-7fcd-498f-8412-563b16e6c986 req-8d70cdb6-1137-4bc8-a75f-83bce0c681ed service nova] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.143398] env[63355]: DEBUG nova.scheduler.client.report [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.291395] env[63355]: INFO nova.compute.manager [-] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Took 1.02 seconds to deallocate network for instance. [ 724.293788] env[63355]: DEBUG nova.compute.claims [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 724.294051] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.313765] env[63355]: DEBUG oslo_concurrency.lockutils [req-1b0a08b2-7fcd-498f-8412-563b16e6c986 req-8d70cdb6-1137-4bc8-a75f-83bce0c681ed service nova] Releasing lock "refresh_cache-9478ce7a-99c6-4e95-82b0-f8a71ce94a90" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.314613] env[63355]: DEBUG nova.compute.manager [req-1b0a08b2-7fcd-498f-8412-563b16e6c986 req-8d70cdb6-1137-4bc8-a75f-83bce0c681ed service nova] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Received event network-vif-deleted-6973d5f6-e86d-4734-85ff-ff458f776431 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.318982] env[63355]: ERROR nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ac49af64-6ba9-4980-b8dc-ac660a40adae, please check neutron logs for more information. [ 724.318982] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 724.318982] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.318982] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 724.318982] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.318982] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 724.318982] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.318982] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 724.318982] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.318982] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 724.318982] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.318982] env[63355]: ERROR nova.compute.manager raise self.value [ 724.318982] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.318982] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 724.318982] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.318982] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 724.319408] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.319408] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 724.319408] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ac49af64-6ba9-4980-b8dc-ac660a40adae, please check neutron logs for more information. [ 724.319408] env[63355]: ERROR nova.compute.manager [ 724.319408] env[63355]: Traceback (most recent call last): [ 724.319408] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 724.319408] env[63355]: listener.cb(fileno) [ 724.319408] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.319408] env[63355]: result = function(*args, **kwargs) [ 724.319408] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.319408] env[63355]: return func(*args, **kwargs) [ 724.319408] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.319408] env[63355]: raise e [ 724.319408] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.319408] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 724.319408] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.319408] env[63355]: created_port_ids = self._update_ports_for_instance( [ 724.319408] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.319408] env[63355]: with excutils.save_and_reraise_exception(): [ 724.319408] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.319408] env[63355]: self.force_reraise() [ 724.319408] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.319408] env[63355]: raise self.value [ 724.319408] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.319408] env[63355]: updated_port = self._update_port( [ 724.319408] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.319408] env[63355]: _ensure_no_port_binding_failure(port) [ 724.319408] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.319408] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 724.320135] env[63355]: nova.exception.PortBindingFailed: Binding failed for port ac49af64-6ba9-4980-b8dc-ac660a40adae, please check neutron logs for more information. [ 724.320135] env[63355]: Removing descriptor: 17 [ 724.649198] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.016s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.650459] env[63355]: ERROR nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6b105f42-97df-4f8b-a719-3fcd47feeb49, please check neutron logs for more information. [ 724.650459] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Traceback (most recent call last): [ 724.650459] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.650459] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] self.driver.spawn(context, instance, image_meta, [ 724.650459] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 724.650459] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.650459] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.650459] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] vm_ref = self.build_virtual_machine(instance, [ 724.650459] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.650459] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.650459] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] for vif in network_info: [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] return self._sync_wrapper(fn, *args, **kwargs) [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] self.wait() [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] self[:] = self._gt.wait() [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] return self._exit_event.wait() [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] result = hub.switch() [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 724.650771] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] return self.greenlet.switch() [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] result = function(*args, **kwargs) [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] return func(*args, **kwargs) [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] raise e [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] nwinfo = self.network_api.allocate_for_instance( [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] created_port_ids = self._update_ports_for_instance( [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] with excutils.save_and_reraise_exception(): [ 724.651088] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] self.force_reraise() [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] raise self.value [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] updated_port = self._update_port( [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] _ensure_no_port_binding_failure(port) [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] raise exception.PortBindingFailed(port_id=port['id']) [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] nova.exception.PortBindingFailed: Binding failed for port 6b105f42-97df-4f8b-a719-3fcd47feeb49, please check neutron logs for more information. [ 724.651408] env[63355]: ERROR nova.compute.manager [instance: 617b624f-b978-4f98-bbbc-befa006334e3] [ 724.651699] env[63355]: DEBUG nova.compute.utils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Binding failed for port 6b105f42-97df-4f8b-a719-3fcd47feeb49, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 724.651699] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.745s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.653026] env[63355]: INFO nova.compute.claims [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.656107] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Build of instance 617b624f-b978-4f98-bbbc-befa006334e3 was re-scheduled: Binding failed for port 6b105f42-97df-4f8b-a719-3fcd47feeb49, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 724.656558] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 724.656783] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.656930] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquired lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.657100] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.658563] env[63355]: DEBUG nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.687413] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.687660] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.687813] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.687992] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.688151] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.689618] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.689618] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.689618] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.689618] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.689618] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.689783] env[63355]: DEBUG nova.virt.hardware [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.689939] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a4421b-1ea9-4d3f-8064-fcf3134950f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.699256] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01059fe9-ae82-4765-88e0-59f544d6379c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.713442] env[63355]: ERROR nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ac49af64-6ba9-4980-b8dc-ac660a40adae, please check neutron logs for more information. [ 724.713442] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Traceback (most recent call last): [ 724.713442] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 724.713442] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] yield resources [ 724.713442] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.713442] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] self.driver.spawn(context, instance, image_meta, [ 724.713442] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 724.713442] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.713442] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.713442] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] vm_ref = self.build_virtual_machine(instance, [ 724.713442] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] for vif in network_info: [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] return self._sync_wrapper(fn, *args, **kwargs) [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] self.wait() [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] self[:] = self._gt.wait() [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] return self._exit_event.wait() [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.713863] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] current.throw(*self._exc) [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] result = function(*args, **kwargs) [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] return func(*args, **kwargs) [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] raise e [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] nwinfo = self.network_api.allocate_for_instance( [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] created_port_ids = self._update_ports_for_instance( [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] with excutils.save_and_reraise_exception(): [ 724.714284] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] self.force_reraise() [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] raise self.value [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] updated_port = self._update_port( [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] _ensure_no_port_binding_failure(port) [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] raise exception.PortBindingFailed(port_id=port['id']) [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] nova.exception.PortBindingFailed: Binding failed for port ac49af64-6ba9-4980-b8dc-ac660a40adae, please check neutron logs for more information. [ 724.714781] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] [ 724.714781] env[63355]: INFO nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Terminating instance [ 724.717737] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "refresh_cache-b7dcc653-a430-48c4-9185-3e8200cf1005" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.717893] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "refresh_cache-b7dcc653-a430-48c4-9185-3e8200cf1005" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.718075] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.179141] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.236905] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.252869] env[63355]: DEBUG nova.compute.manager [req-4d4317bd-e27d-49de-a01f-d8a3a43e30fa req-08c6eb1a-0054-4e9d-881b-b5a0c0ae1db6 service nova] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Received event network-changed-ac49af64-6ba9-4980-b8dc-ac660a40adae {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.253074] env[63355]: DEBUG nova.compute.manager [req-4d4317bd-e27d-49de-a01f-d8a3a43e30fa req-08c6eb1a-0054-4e9d-881b-b5a0c0ae1db6 service nova] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Refreshing instance network info cache due to event network-changed-ac49af64-6ba9-4980-b8dc-ac660a40adae. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.253259] env[63355]: DEBUG oslo_concurrency.lockutils [req-4d4317bd-e27d-49de-a01f-d8a3a43e30fa req-08c6eb1a-0054-4e9d-881b-b5a0c0ae1db6 service nova] Acquiring lock "refresh_cache-b7dcc653-a430-48c4-9185-3e8200cf1005" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.262538] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.338989] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.766706] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Releasing lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.766706] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 725.766706] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.766706] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.780734] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.844695] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "refresh_cache-b7dcc653-a430-48c4-9185-3e8200cf1005" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.844695] env[63355]: DEBUG nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.844695] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 725.844695] env[63355]: DEBUG oslo_concurrency.lockutils [req-4d4317bd-e27d-49de-a01f-d8a3a43e30fa req-08c6eb1a-0054-4e9d-881b-b5a0c0ae1db6 service nova] Acquired lock "refresh_cache-b7dcc653-a430-48c4-9185-3e8200cf1005" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.844900] env[63355]: DEBUG nova.network.neutron [req-4d4317bd-e27d-49de-a01f-d8a3a43e30fa req-08c6eb1a-0054-4e9d-881b-b5a0c0ae1db6 service nova] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Refreshing network info cache for port ac49af64-6ba9-4980-b8dc-ac660a40adae {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 725.846391] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ea7f321-8e07-48ba-851b-34e1d7712c90 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.855346] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28508794-736f-47f8-8584-51efe431cfc0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.880698] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b7dcc653-a430-48c4-9185-3e8200cf1005 could not be found. [ 725.880925] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 725.881120] env[63355]: INFO nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Took 0.04 seconds to destroy the instance on the hypervisor. [ 725.881371] env[63355]: DEBUG oslo.service.loopingcall [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.883831] env[63355]: DEBUG nova.compute.manager [-] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.883925] env[63355]: DEBUG nova.network.neutron [-] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.900255] env[63355]: DEBUG nova.network.neutron [-] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.099189] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aab6eab-c454-4da8-9f09-dd90256005d9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.107023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fac7d24-4f09-4a7f-bec8-bd820b5272b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.137382] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c2ad40-23f1-46ae-a880-75fcdd8e3eda {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.144824] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c113e7c1-90b7-42a6-91e2-d90d71364327 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.157631] env[63355]: DEBUG nova.compute.provider_tree [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.284684] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.384263] env[63355]: DEBUG nova.network.neutron [req-4d4317bd-e27d-49de-a01f-d8a3a43e30fa req-08c6eb1a-0054-4e9d-881b-b5a0c0ae1db6 service nova] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.402280] env[63355]: DEBUG nova.network.neutron [-] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.479952] env[63355]: DEBUG nova.network.neutron [req-4d4317bd-e27d-49de-a01f-d8a3a43e30fa req-08c6eb1a-0054-4e9d-881b-b5a0c0ae1db6 service nova] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.660489] env[63355]: DEBUG nova.scheduler.client.report [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.787787] env[63355]: INFO nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Took 1.02 seconds to deallocate network for instance. [ 726.907198] env[63355]: INFO nova.compute.manager [-] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Took 1.02 seconds to deallocate network for instance. [ 726.909608] env[63355]: DEBUG nova.compute.claims [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 726.909796] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.982325] env[63355]: DEBUG oslo_concurrency.lockutils [req-4d4317bd-e27d-49de-a01f-d8a3a43e30fa req-08c6eb1a-0054-4e9d-881b-b5a0c0ae1db6 service nova] Releasing lock "refresh_cache-b7dcc653-a430-48c4-9185-3e8200cf1005" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.982589] env[63355]: DEBUG nova.compute.manager [req-4d4317bd-e27d-49de-a01f-d8a3a43e30fa req-08c6eb1a-0054-4e9d-881b-b5a0c0ae1db6 service nova] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Received event network-vif-deleted-ac49af64-6ba9-4980-b8dc-ac660a40adae {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.165767] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.166318] env[63355]: DEBUG nova.compute.manager [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.168927] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.237s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.170337] env[63355]: INFO nova.compute.claims [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.675183] env[63355]: DEBUG nova.compute.utils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.678751] env[63355]: DEBUG nova.compute.manager [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Not allocating networking since 'none' was specified. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 727.818411] env[63355]: INFO nova.scheduler.client.report [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Deleted allocations for instance 617b624f-b978-4f98-bbbc-befa006334e3 [ 728.179247] env[63355]: DEBUG nova.compute.manager [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 728.325773] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "617b624f-b978-4f98-bbbc-befa006334e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.229s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.327179] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "617b624f-b978-4f98-bbbc-befa006334e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 8.623s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.328203] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "617b624f-b978-4f98-bbbc-befa006334e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.328203] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "617b624f-b978-4f98-bbbc-befa006334e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.328203] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "617b624f-b978-4f98-bbbc-befa006334e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.329718] env[63355]: INFO nova.compute.manager [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Terminating instance [ 728.331449] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.331531] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquired lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.331727] env[63355]: DEBUG nova.network.neutron [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.562241] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52d2987-8aea-4648-a8a4-a6d64e328282 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.569569] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7a11f9-5e44-46eb-9286-bb239f09f17d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.600387] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fc6548-c854-4968-b0c4-3b0c2544ebea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.607776] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15f998a-7dc0-4035-91fb-6d93b2cab0ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.620621] env[63355]: DEBUG nova.compute.provider_tree [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.830185] env[63355]: DEBUG nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.860397] env[63355]: DEBUG nova.network.neutron [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.978998] env[63355]: DEBUG nova.network.neutron [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.123714] env[63355]: DEBUG nova.scheduler.client.report [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.190829] env[63355]: DEBUG nova.compute.manager [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.213937] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.214251] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.214435] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.214626] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.214771] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.214917] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.215143] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.215296] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.215457] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.215612] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.215776] env[63355]: DEBUG nova.virt.hardware [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.216646] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3c381b-d92d-4dd3-9b6d-cc14239f3a89 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.226395] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276df490-9f62-4fd2-a940-6433d870fd97 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.239498] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 729.245074] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Creating folder: Project (46dde55450244b87931f62b4c09f597b). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 729.245336] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14342a4d-fa33-484f-aa7b-ab08143d0430 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.255144] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Created folder: Project (46dde55450244b87931f62b4c09f597b) in parent group-v287607. [ 729.255322] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Creating folder: Instances. Parent ref: group-v287621. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 729.255529] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2680ad2c-4820-4beb-8dac-2a8ec72d5cb1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.263513] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Created folder: Instances in parent group-v287621. [ 729.263710] env[63355]: DEBUG oslo.service.loopingcall [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.263883] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 729.264106] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ec3f384-69ee-47e1-9eef-5bdac581fd5d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.279751] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 729.279751] env[63355]: value = "task-1349523" [ 729.279751] env[63355]: _type = "Task" [ 729.279751] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.286530] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349523, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.353652] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.481650] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Releasing lock "refresh_cache-617b624f-b978-4f98-bbbc-befa006334e3" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.482099] env[63355]: DEBUG nova.compute.manager [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 729.482350] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 729.482667] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f8ae5d9c-2d02-463f-8a0c-50b63f452427 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.491907] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45874eae-51d2-45a7-beb1-409833151a4a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.515219] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 617b624f-b978-4f98-bbbc-befa006334e3 could not be found. [ 729.515219] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 729.515219] env[63355]: INFO nova.compute.manager [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 729.515219] env[63355]: DEBUG oslo.service.loopingcall [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.515451] env[63355]: DEBUG nova.compute.manager [-] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.515451] env[63355]: DEBUG nova.network.neutron [-] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.542955] env[63355]: DEBUG nova.network.neutron [-] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.628863] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.629054] env[63355]: DEBUG nova.compute.manager [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.631568] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.617s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.789898] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349523, 'name': CreateVM_Task, 'duration_secs': 0.264763} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.790260] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 729.790810] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.791115] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.791539] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 729.791892] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdc51561-f822-4236-aae0-3fe92d4a46e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.796498] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for the task: (returnval){ [ 729.796498] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52378da3-60a3-3ca1-b6aa-850e79320ce3" [ 729.796498] env[63355]: _type = "Task" [ 729.796498] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.805213] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52378da3-60a3-3ca1-b6aa-850e79320ce3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.045818] env[63355]: DEBUG nova.network.neutron [-] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.134981] env[63355]: DEBUG nova.compute.utils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.139789] env[63355]: DEBUG nova.compute.manager [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Not allocating networking since 'none' was specified. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 730.307112] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52378da3-60a3-3ca1-b6aa-850e79320ce3, 'name': SearchDatastore_Task, 'duration_secs': 0.010177} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.307397] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.307621] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 730.307884] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.307987] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.308177] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 730.308426] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc2d1a83-1247-4b2e-9b89-61c29072ee1a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.315920] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 730.316096] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 730.316749] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9acf7025-e2d7-419f-ba86-fa4e52b05c46 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.321499] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for the task: (returnval){ [ 730.321499] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521d9c35-1a69-1ed0-4011-8191c2540425" [ 730.321499] env[63355]: _type = "Task" [ 730.321499] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.328275] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521d9c35-1a69-1ed0-4011-8191c2540425, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.548570] env[63355]: INFO nova.compute.manager [-] [instance: 617b624f-b978-4f98-bbbc-befa006334e3] Took 1.03 seconds to deallocate network for instance. [ 730.640718] env[63355]: DEBUG nova.compute.manager [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 730.663932] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 12f9b2b4-3c78-43c2-bb9b-d44b81c83123 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 730.664173] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance ad05eda3-efd7-4040-be9c-6a47df02889b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 730.664245] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 9478ce7a-99c6-4e95-82b0-f8a71ce94a90 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 730.664362] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b7dcc653-a430-48c4-9185-3e8200cf1005 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 730.664476] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b2bf4912-6b88-4ece-95c9-e9fd1cf906df actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 730.664588] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance e5c80c05-0e48-415a-ac91-a53e1c8819f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 730.832069] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521d9c35-1a69-1ed0-4011-8191c2540425, 'name': SearchDatastore_Task, 'duration_secs': 0.008109} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.832843] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf053ce4-0a1f-46e1-ad50-5f6c3658b186 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.837982] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for the task: (returnval){ [ 730.837982] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f6a742-02f6-1372-ed09-19aaff4d2b59" [ 730.837982] env[63355]: _type = "Task" [ 730.837982] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.845388] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f6a742-02f6-1372-ed09-19aaff4d2b59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.167882] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 9dd8b658-886a-4f11-ab9b-059c4265436c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 731.352193] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f6a742-02f6-1372-ed09-19aaff4d2b59, 'name': SearchDatastore_Task, 'duration_secs': 0.008965} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.352468] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.352720] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] b2bf4912-6b88-4ece-95c9-e9fd1cf906df/b2bf4912-6b88-4ece-95c9-e9fd1cf906df.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 731.352978] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ef668d2a-401b-454e-82ff-6d240c29dcf8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.359974] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for the task: (returnval){ [ 731.359974] env[63355]: value = "task-1349524" [ 731.359974] env[63355]: _type = "Task" [ 731.359974] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.368334] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349524, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.573052] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62510497-ee37-405e-a142-47dd910d26d8 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "617b624f-b978-4f98-bbbc-befa006334e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.246s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.652486] env[63355]: DEBUG nova.compute.manager [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 731.670478] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance d7589681-fe6c-4dc6-b533-ade8ae634d58 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 731.678906] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 731.679170] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 731.679375] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.679599] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 731.679791] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.679946] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 731.680278] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 731.680337] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 731.680504] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 731.680664] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 731.680838] env[63355]: DEBUG nova.virt.hardware [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 731.681720] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c00507-2d2b-4b4b-ba6a-7a096cdf2653 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.692128] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f3e01dc-aa9e-4fbf-88d2-4413fde99925 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.708795] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 731.714624] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Creating folder: Project (dd22f30127a8424ca093e65e87ff64f1). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 731.715494] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53ffe471-636d-4322-9caa-d8a35d712eb0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.727788] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Created folder: Project (dd22f30127a8424ca093e65e87ff64f1) in parent group-v287607. [ 731.728007] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Creating folder: Instances. Parent ref: group-v287624. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 731.728262] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c750c542-0b0c-4a7a-9c47-3b6ef60a94cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.738097] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Created folder: Instances in parent group-v287624. [ 731.738363] env[63355]: DEBUG oslo.service.loopingcall [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.738562] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 731.738771] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8906414-6642-445d-b2da-d8a3c4ea0845 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.756951] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 731.756951] env[63355]: value = "task-1349527" [ 731.756951] env[63355]: _type = "Task" [ 731.756951] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.766681] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349527, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.869612] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349524, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476653} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.869886] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] b2bf4912-6b88-4ece-95c9-e9fd1cf906df/b2bf4912-6b88-4ece-95c9-e9fd1cf906df.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 731.870110] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 731.870369] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03e859fb-e3ab-4e02-8b0d-1d9e0ffb07bd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.875882] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for the task: (returnval){ [ 731.875882] env[63355]: value = "task-1349528" [ 731.875882] env[63355]: _type = "Task" [ 731.875882] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.883148] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349528, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.177409] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 17a904fe-ab4b-46ed-98c8-175987710643 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 732.268842] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349527, 'name': CreateVM_Task, 'duration_secs': 0.248637} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.268842] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 732.268842] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.268974] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.269231] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 732.269474] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2c3578a-b2f2-41c9-afeb-78d42c374e3c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.273847] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 732.273847] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d58727-c299-2de2-a085-e34e8a348541" [ 732.273847] env[63355]: _type = "Task" [ 732.273847] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.281216] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d58727-c299-2de2-a085-e34e8a348541, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.386033] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349528, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067196} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.386311] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 732.387060] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bba37a-d2c2-4e03-aaa3-1e60b500ef51 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.406020] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] b2bf4912-6b88-4ece-95c9-e9fd1cf906df/b2bf4912-6b88-4ece-95c9-e9fd1cf906df.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 732.406267] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94b1c4d4-81d3-4a46-9cd9-1daac2e0b04b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.424802] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for the task: (returnval){ [ 732.424802] env[63355]: value = "task-1349529" [ 732.424802] env[63355]: _type = "Task" [ 732.424802] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.433534] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349529, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.679914] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 732.784579] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d58727-c299-2de2-a085-e34e8a348541, 'name': SearchDatastore_Task, 'duration_secs': 0.008653} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.784880] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.785182] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 732.785472] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.785618] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.785797] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 732.786153] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b13ef3af-f057-4ad0-95a0-15d68ce4a0cb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.794337] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 732.794505] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 732.795205] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be77b145-ab04-4968-af96-331e9c460a34 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.800288] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 732.800288] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a4682f-1704-fda9-e2e8-4f45843a26f0" [ 732.800288] env[63355]: _type = "Task" [ 732.800288] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.807533] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a4682f-1704-fda9-e2e8-4f45843a26f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.934164] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349529, 'name': ReconfigVM_Task, 'duration_secs': 0.262718} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.934470] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Reconfigured VM instance instance-00000021 to attach disk [datastore2] b2bf4912-6b88-4ece-95c9-e9fd1cf906df/b2bf4912-6b88-4ece-95c9-e9fd1cf906df.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 732.935081] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5a43e18-157e-4142-8715-aaeb9565af3a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.941581] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for the task: (returnval){ [ 732.941581] env[63355]: value = "task-1349530" [ 732.941581] env[63355]: _type = "Task" [ 732.941581] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.949826] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349530, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.182923] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 733.310309] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a4682f-1704-fda9-e2e8-4f45843a26f0, 'name': SearchDatastore_Task, 'duration_secs': 0.008679} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.311076] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab8a54cb-f5ec-460c-9886-95fdc904cea6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.316253] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 733.316253] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52656e9f-6289-f8ab-e2ff-f04d1be5a5d8" [ 733.316253] env[63355]: _type = "Task" [ 733.316253] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.323864] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52656e9f-6289-f8ab-e2ff-f04d1be5a5d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.452022] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349530, 'name': Rename_Task, 'duration_secs': 0.124663} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.452022] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 733.452301] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b55ea8a-1e7f-4c29-acb1-669c288dbbbb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.458509] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for the task: (returnval){ [ 733.458509] env[63355]: value = "task-1349531" [ 733.458509] env[63355]: _type = "Task" [ 733.458509] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.465723] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349531, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.685975] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 1446c7da-415f-43ee-9ff1-2266bd3038e4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 733.827338] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52656e9f-6289-f8ab-e2ff-f04d1be5a5d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009704} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.827912] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.828200] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] e5c80c05-0e48-415a-ac91-a53e1c8819f0/e5c80c05-0e48-415a-ac91-a53e1c8819f0.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 733.828463] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51b66a2f-6b08-4085-8c2c-9d3ac84daa13 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.834654] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 733.834654] env[63355]: value = "task-1349532" [ 733.834654] env[63355]: _type = "Task" [ 733.834654] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.842375] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349532, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.967561] env[63355]: DEBUG oslo_vmware.api [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349531, 'name': PowerOnVM_Task, 'duration_secs': 0.399515} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.967823] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 733.968042] env[63355]: INFO nova.compute.manager [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Took 4.78 seconds to spawn the instance on the hypervisor. [ 733.968684] env[63355]: DEBUG nova.compute.manager [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 733.968966] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a71bf81-9a55-4f72-879a-628a046942f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.189530] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance ec4726b2-38f9-4499-a5b0-a70db0d697f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 734.345144] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349532, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467599} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.346121] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] e5c80c05-0e48-415a-ac91-a53e1c8819f0/e5c80c05-0e48-415a-ac91-a53e1c8819f0.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 734.346121] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.346121] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ccc67dc-f68e-47e6-91e3-11bff3e4c83a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.352091] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 734.352091] env[63355]: value = "task-1349533" [ 734.352091] env[63355]: _type = "Task" [ 734.352091] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.359834] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349533, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.488443] env[63355]: DEBUG nova.compute.manager [None req-ffb643cb-26a5-440e-a7d1-0cbb8881b9d5 tempest-ServerDiagnosticsV248Test-904048433 tempest-ServerDiagnosticsV248Test-904048433-project-admin] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 734.488860] env[63355]: INFO nova.compute.manager [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Took 27.61 seconds to build instance. [ 734.491019] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381ef5df-a94c-40d1-8562-f9a5198040be {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.497861] env[63355]: INFO nova.compute.manager [None req-ffb643cb-26a5-440e-a7d1-0cbb8881b9d5 tempest-ServerDiagnosticsV248Test-904048433 tempest-ServerDiagnosticsV248Test-904048433-project-admin] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Retrieving diagnostics [ 734.498844] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92536bb8-1deb-4a0f-913f-04e44e0e3f4d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.693808] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 3c04c546-a9fa-4cbd-98d4-3ea92918a918 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 734.862643] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349533, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065511} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.864819] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 734.864819] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e542ff4-a2e7-49cb-a173-001459c82ff0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.884330] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] e5c80c05-0e48-415a-ac91-a53e1c8819f0/e5c80c05-0e48-415a-ac91-a53e1c8819f0.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 734.884638] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8880d531-c26d-4353-8d95-7d42b655774b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.904364] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 734.904364] env[63355]: value = "task-1349534" [ 734.904364] env[63355]: _type = "Task" [ 734.904364] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.914593] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349534, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.990923] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e83f6e71-ad0e-46e5-876b-4d2641d15f50 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lock "b2bf4912-6b88-4ece-95c9-e9fd1cf906df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.920s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.198111] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 735.414770] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349534, 'name': ReconfigVM_Task, 'duration_secs': 0.253514} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.415065] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Reconfigured VM instance instance-00000022 to attach disk [datastore2] e5c80c05-0e48-415a-ac91-a53e1c8819f0/e5c80c05-0e48-415a-ac91-a53e1c8819f0.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 735.416053] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-baea7d64-7933-474a-945e-dc04b148061e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.423032] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 735.423032] env[63355]: value = "task-1349535" [ 735.423032] env[63355]: _type = "Task" [ 735.423032] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.431008] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349535, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.494059] env[63355]: DEBUG nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 735.700301] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 282f1b55-8a74-467a-9ec8-0684eb780970 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 735.935102] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349535, 'name': Rename_Task, 'duration_secs': 0.13071} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.935426] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 735.935673] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-321b6f88-f85f-4ff1-8a0f-2cc0fdfc21a6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.944213] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 735.944213] env[63355]: value = "task-1349536" [ 735.944213] env[63355]: _type = "Task" [ 735.944213] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.952420] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349536, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.022499] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.203451] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 736.454625] env[63355]: DEBUG oslo_vmware.api [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349536, 'name': PowerOnVM_Task, 'duration_secs': 0.395716} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.455055] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 736.455384] env[63355]: INFO nova.compute.manager [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Took 4.80 seconds to spawn the instance on the hypervisor. [ 736.455649] env[63355]: DEBUG nova.compute.manager [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 736.456491] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35e57e9-0409-4cdb-a90d-12289324dc30 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.706605] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 795b61f6-a0f3-4aab-94b3-2540ccbb7312 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 736.972572] env[63355]: INFO nova.compute.manager [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Took 28.06 seconds to build instance. [ 737.210022] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7816d808-c9dd-403f-b993-c6d9dc458c7b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 737.475344] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f922a04-d1e1-462d-9b3d-101ae330c325 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "e5c80c05-0e48-415a-ac91-a53e1c8819f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.500s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.714330] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance a1e1e0ea-debd-4529-b175-4bdf44e4de31 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 737.977538] env[63355]: DEBUG nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.028944] env[63355]: INFO nova.compute.manager [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Rebuilding instance [ 738.076016] env[63355]: DEBUG nova.compute.manager [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 738.076881] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5360c49-07ca-410f-8de6-17e93a04f958 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.216712] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 21e17d32-40db-4dc1-9e73-fa6d35a4c06a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 738.499150] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.587542] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 738.587852] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbc2cdac-0873-475f-8a46-46dd23bdd728 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.595264] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 738.595264] env[63355]: value = "task-1349537" [ 738.595264] env[63355]: _type = "Task" [ 738.595264] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.603139] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.721074] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7d79c9cf-66a9-4117-b090-0ab0676c3114 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 739.104537] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349537, 'name': PowerOffVM_Task, 'duration_secs': 0.119271} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.104830] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 739.105093] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.105918] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3eeb43e-bfec-4694-bc82-79b1d9f6ebc5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.112732] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 739.112952] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2085b9d1-8fda-432e-9c93-5aa786c62a0c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.137662] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 739.137887] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 739.138081] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Deleting the datastore file [datastore2] e5c80c05-0e48-415a-ac91-a53e1c8819f0 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 739.138332] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d6108eb-cdeb-4f06-b40c-19f8f379a91b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.144449] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 739.144449] env[63355]: value = "task-1349539" [ 739.144449] env[63355]: _type = "Task" [ 739.144449] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.152023] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349539, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.223922] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 4ded6053-9afc-444e-bdfc-244265ed19a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 739.654451] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349539, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106108} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.654694] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 739.654898] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 739.655051] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.727493] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance dd80c3be-ec92-4a67-a373-3e2276752feb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 740.230104] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 9989c129-07d3-4af9-9a86-1e2746d6ac70 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 740.686636] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.686878] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.687041] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.687227] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.687371] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.687513] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.687718] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.687871] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.688137] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.688306] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.688472] env[63355]: DEBUG nova.virt.hardware [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.689336] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca101302-7668-4808-a88e-2deacbc3113c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.697400] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03043c4a-3905-4184-af7c-1f8e63446673 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.710995] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 740.715865] env[63355]: DEBUG oslo.service.loopingcall [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.716126] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 740.716334] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31ac31b4-7edb-4634-863a-6aff7593b379 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.733038] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8a44bcfb-406c-45f6-b027-86d8980bc5d1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 740.734609] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.734609] env[63355]: value = "task-1349540" [ 740.734609] env[63355]: _type = "Task" [ 740.734609] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.742566] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349540, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.236017] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance e49ba26c-486c-43aa-9595-231e7f31a522 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 741.247356] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349540, 'name': CreateVM_Task, 'duration_secs': 0.248128} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.247526] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 741.248279] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.248666] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.248761] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 741.249701] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b4b5522-c97d-4f7e-8edd-bf0c5a3de923 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.254654] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 741.254654] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5230d3fa-f32c-d413-8f2f-103d2e8f85bf" [ 741.254654] env[63355]: _type = "Task" [ 741.254654] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.261901] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5230d3fa-f32c-d413-8f2f-103d2e8f85bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.742965] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7dd28cb7-d249-4b8e-beb3-c959f15e912f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 741.765961] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5230d3fa-f32c-d413-8f2f-103d2e8f85bf, 'name': SearchDatastore_Task, 'duration_secs': 0.008504} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.766404] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.766544] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 741.766811] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.766909] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.767102] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 741.767411] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a9308ea-06dc-4830-85d1-7de742450501 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.776189] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 741.776367] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 741.777113] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcda6338-7617-4442-8949-7f75fcfa93b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.781950] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 741.781950] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52906a9b-e660-f6ce-9836-6056248773b5" [ 741.781950] env[63355]: _type = "Task" [ 741.781950] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.789162] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52906a9b-e660-f6ce-9836-6056248773b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.246581] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 947a33d3-88c1-45ae-805c-eda2bc7626f1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 742.292632] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52906a9b-e660-f6ce-9836-6056248773b5, 'name': SearchDatastore_Task, 'duration_secs': 0.007543} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.293417] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ece451ec-84b8-42b6-b640-39e15d14342e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.298385] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 742.298385] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5296545a-6bf2-3481-e329-75db5d46476c" [ 742.298385] env[63355]: _type = "Task" [ 742.298385] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.306861] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5296545a-6bf2-3481-e329-75db5d46476c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.749036] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 3d918cab-6ed6-4a37-a024-28e3db1b779c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 742.808425] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5296545a-6bf2-3481-e329-75db5d46476c, 'name': SearchDatastore_Task, 'duration_secs': 0.008565} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.808661] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.808915] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] e5c80c05-0e48-415a-ac91-a53e1c8819f0/e5c80c05-0e48-415a-ac91-a53e1c8819f0.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 742.809185] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15f93f50-77f5-489f-85ea-d5b062b75b01 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.815319] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 742.815319] env[63355]: value = "task-1349541" [ 742.815319] env[63355]: _type = "Task" [ 742.815319] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.822126] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349541, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.252249] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 5591bb55-83d7-4301-a3f9-fde945632344 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 743.324306] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349541, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487547} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.324579] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] e5c80c05-0e48-415a-ac91-a53e1c8819f0/e5c80c05-0e48-415a-ac91-a53e1c8819f0.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 743.324786] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 743.325059] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d35ab38-a4af-49d0-8299-44921d7e792a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.332141] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 743.332141] env[63355]: value = "task-1349542" [ 743.332141] env[63355]: _type = "Task" [ 743.332141] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.339437] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349542, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.755284] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 386e847e-967b-4247-9730-cdc5ac251474 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 743.841095] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349542, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060783} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.841587] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 743.842475] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076e728a-f6a4-4c47-a826-4ff6d9662d52 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.861149] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] e5c80c05-0e48-415a-ac91-a53e1c8819f0/e5c80c05-0e48-415a-ac91-a53e1c8819f0.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 743.862049] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b558d05b-6e24-4585-aac1-f33eb764e541 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.879564] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 743.879564] env[63355]: value = "task-1349543" [ 743.879564] env[63355]: _type = "Task" [ 743.879564] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.886560] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349543, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.259124] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 0b660647-697a-445b-bca4-82ef6975ea75 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 744.389678] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349543, 'name': ReconfigVM_Task, 'duration_secs': 0.254814} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.389969] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Reconfigured VM instance instance-00000022 to attach disk [datastore1] e5c80c05-0e48-415a-ac91-a53e1c8819f0/e5c80c05-0e48-415a-ac91-a53e1c8819f0.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 744.390791] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d9d1e98-00a9-471c-a8c5-e5ea1b363c84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.396884] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 744.396884] env[63355]: value = "task-1349544" [ 744.396884] env[63355]: _type = "Task" [ 744.396884] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.404636] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349544, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.762863] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 744.764826] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 744.764826] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 744.909600] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349544, 'name': Rename_Task, 'duration_secs': 0.127111} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.909877] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 744.910131] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3e46665-3dc5-42fd-a950-caf119eb4c00 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.918327] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 744.918327] env[63355]: value = "task-1349545" [ 744.918327] env[63355]: _type = "Task" [ 744.918327] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.927518] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349545, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.029521] env[63355]: DEBUG nova.compute.manager [None req-c36878d4-4e67-481f-a5d6-16d654d37db8 tempest-ServerDiagnosticsV248Test-904048433 tempest-ServerDiagnosticsV248Test-904048433-project-admin] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 745.030758] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc54c0f-b227-427b-8485-91ee361a6c14 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.038806] env[63355]: INFO nova.compute.manager [None req-c36878d4-4e67-481f-a5d6-16d654d37db8 tempest-ServerDiagnosticsV248Test-904048433 tempest-ServerDiagnosticsV248Test-904048433-project-admin] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Retrieving diagnostics [ 745.041507] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5b9704-b4df-4a08-8c70-c4ce165da195 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.144096] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43d94aa-eafc-4fbb-8720-875d85e11c54 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.151241] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee84981-6b8f-4194-babb-a70500b6581b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.180021] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144d5cdb-bf7a-47a9-8d46-e4d13cc2a137 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.187477] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10db55a4-6d96-4aeb-962c-a33a2e22767a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.202618] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.428747] env[63355]: DEBUG oslo_vmware.api [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349545, 'name': PowerOnVM_Task, 'duration_secs': 0.407669} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.429039] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 745.429242] env[63355]: DEBUG nova.compute.manager [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 745.429979] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f31e97-0c49-4455-9a12-036a79538240 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.704864] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.948481] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.193961] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquiring lock "b2bf4912-6b88-4ece-95c9-e9fd1cf906df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.194302] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lock "b2bf4912-6b88-4ece-95c9-e9fd1cf906df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.194538] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquiring lock "b2bf4912-6b88-4ece-95c9-e9fd1cf906df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.194782] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lock "b2bf4912-6b88-4ece-95c9-e9fd1cf906df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.194881] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lock "b2bf4912-6b88-4ece-95c9-e9fd1cf906df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.197164] env[63355]: INFO nova.compute.manager [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Terminating instance [ 746.199685] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquiring lock "refresh_cache-b2bf4912-6b88-4ece-95c9-e9fd1cf906df" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.199841] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquired lock "refresh_cache-b2bf4912-6b88-4ece-95c9-e9fd1cf906df" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.200010] env[63355]: DEBUG nova.network.neutron [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.209438] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 746.209647] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 16.578s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.210133] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.090s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.211642] env[63355]: INFO nova.compute.claims [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.718782] env[63355]: DEBUG nova.network.neutron [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.754304] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "e5c80c05-0e48-415a-ac91-a53e1c8819f0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.754582] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "e5c80c05-0e48-415a-ac91-a53e1c8819f0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.754784] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "e5c80c05-0e48-415a-ac91-a53e1c8819f0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.754954] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "e5c80c05-0e48-415a-ac91-a53e1c8819f0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.755141] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "e5c80c05-0e48-415a-ac91-a53e1c8819f0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.757187] env[63355]: INFO nova.compute.manager [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Terminating instance [ 746.758901] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "refresh_cache-e5c80c05-0e48-415a-ac91-a53e1c8819f0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.759124] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquired lock "refresh_cache-e5c80c05-0e48-415a-ac91-a53e1c8819f0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.759233] env[63355]: DEBUG nova.network.neutron [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.801076] env[63355]: DEBUG nova.network.neutron [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.275741] env[63355]: DEBUG nova.network.neutron [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.302510] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Releasing lock "refresh_cache-b2bf4912-6b88-4ece-95c9-e9fd1cf906df" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.302772] env[63355]: DEBUG nova.compute.manager [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.302971] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.306070] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2754a156-7eb7-47ee-a58d-497f741ef787 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.315311] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 747.315311] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a392ac67-9d3c-4908-849d-4e6eb364237b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.322304] env[63355]: DEBUG oslo_vmware.api [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for the task: (returnval){ [ 747.322304] env[63355]: value = "task-1349546" [ 747.322304] env[63355]: _type = "Task" [ 747.322304] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.331047] env[63355]: DEBUG oslo_vmware.api [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349546, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.333771] env[63355]: DEBUG nova.network.neutron [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.606479] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a0aa0b-6dee-4a72-812e-0bd32c621fb8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.613967] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39091b05-0915-476f-9297-a37a7e5d2fd4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.643483] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ba51f7-aea7-4e26-a382-aedab513db7c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.650414] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24150ef-c75f-44d6-a50f-56a045b1ce9f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.662815] env[63355]: DEBUG nova.compute.provider_tree [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.833092] env[63355]: DEBUG oslo_vmware.api [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349546, 'name': PowerOffVM_Task, 'duration_secs': 0.187592} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.833414] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 747.833528] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 747.833777] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e19ecf3e-bddb-4bf9-beb2-8859f45cf50e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.835965] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Releasing lock "refresh_cache-e5c80c05-0e48-415a-ac91-a53e1c8819f0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.836355] env[63355]: DEBUG nova.compute.manager [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.836546] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.837301] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee870d6-2413-4d55-8e97-f4105aeb89fc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.843756] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 747.843987] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c30c2ab-2771-4599-b4ca-fc700a132680 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.849477] env[63355]: DEBUG oslo_vmware.api [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 747.849477] env[63355]: value = "task-1349548" [ 747.849477] env[63355]: _type = "Task" [ 747.849477] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.859769] env[63355]: DEBUG oslo_vmware.api [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349548, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.861088] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 747.861284] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 747.861495] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Deleting the datastore file [datastore2] b2bf4912-6b88-4ece-95c9-e9fd1cf906df {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 747.861737] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8639dc8e-8117-41b6-a2f9-6b2d76dbedb2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.867830] env[63355]: DEBUG oslo_vmware.api [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for the task: (returnval){ [ 747.867830] env[63355]: value = "task-1349549" [ 747.867830] env[63355]: _type = "Task" [ 747.867830] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.877039] env[63355]: DEBUG oslo_vmware.api [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349549, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.168143] env[63355]: DEBUG nova.scheduler.client.report [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.359711] env[63355]: DEBUG oslo_vmware.api [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349548, 'name': PowerOffVM_Task, 'duration_secs': 0.139106} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.359918] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 748.360177] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 748.360433] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37214a6c-b196-49a6-956a-acd87deb1b14 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.377231] env[63355]: DEBUG oslo_vmware.api [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Task: {'id': task-1349549, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094921} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.377483] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.377665] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 748.377840] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.378017] env[63355]: INFO nova.compute.manager [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Took 1.08 seconds to destroy the instance on the hypervisor. [ 748.378260] env[63355]: DEBUG oslo.service.loopingcall [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.378450] env[63355]: DEBUG nova.compute.manager [-] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.378543] env[63355]: DEBUG nova.network.neutron [-] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.387472] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 748.387702] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 748.387883] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Deleting the datastore file [datastore1] e5c80c05-0e48-415a-ac91-a53e1c8819f0 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 748.388162] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd8fa02b-f268-47be-9824-038cc194192f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.393871] env[63355]: DEBUG oslo_vmware.api [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for the task: (returnval){ [ 748.393871] env[63355]: value = "task-1349551" [ 748.393871] env[63355]: _type = "Task" [ 748.393871] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.401409] env[63355]: DEBUG oslo_vmware.api [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349551, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.402240] env[63355]: DEBUG nova.network.neutron [-] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.671124] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.671663] env[63355]: DEBUG nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.674578] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 34.342s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.904261] env[63355]: DEBUG nova.network.neutron [-] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.905469] env[63355]: DEBUG oslo_vmware.api [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Task: {'id': task-1349551, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09473} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.906063] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.906160] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 748.906259] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.906428] env[63355]: INFO nova.compute.manager [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Took 1.07 seconds to destroy the instance on the hypervisor. [ 748.906659] env[63355]: DEBUG oslo.service.loopingcall [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.906840] env[63355]: DEBUG nova.compute.manager [-] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.906933] env[63355]: DEBUG nova.network.neutron [-] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.938513] env[63355]: DEBUG nova.network.neutron [-] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.179070] env[63355]: DEBUG nova.compute.utils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.184643] env[63355]: DEBUG nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 749.184643] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 749.248868] env[63355]: DEBUG nova.policy [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bea2c2f3fb144cc9adba6ffda59627d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2bcc167f94446c88462022a4a39380e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 749.406680] env[63355]: INFO nova.compute.manager [-] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Took 1.03 seconds to deallocate network for instance. [ 749.440824] env[63355]: DEBUG nova.network.neutron [-] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.600806] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800ce65a-2448-42ca-932f-4f608a7b582a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.611650] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0832427-18b5-4aed-bdf9-d544a0a5d5ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.639635] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Successfully created port: a3482ef0-da11-44f3-84af-7490a7523ddd {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.641961] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de008ab-3a66-404d-bea1-7c2fbcf930dc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.649327] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc74fc79-0a70-47a5-90ac-072d8418d3a0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.662147] env[63355]: DEBUG nova.compute.provider_tree [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.684167] env[63355]: DEBUG nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.914957] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.943577] env[63355]: INFO nova.compute.manager [-] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Took 1.04 seconds to deallocate network for instance. [ 750.167018] env[63355]: DEBUG nova.scheduler.client.report [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.450565] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.518308] env[63355]: DEBUG nova.compute.manager [req-3f3f94de-fb45-4e56-b30a-67769565f895 req-0f1b79d8-7122-49a1-94c3-1c8af279294e service nova] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Received event network-changed-a3482ef0-da11-44f3-84af-7490a7523ddd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.518499] env[63355]: DEBUG nova.compute.manager [req-3f3f94de-fb45-4e56-b30a-67769565f895 req-0f1b79d8-7122-49a1-94c3-1c8af279294e service nova] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Refreshing instance network info cache due to event network-changed-a3482ef0-da11-44f3-84af-7490a7523ddd. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 750.518751] env[63355]: DEBUG oslo_concurrency.lockutils [req-3f3f94de-fb45-4e56-b30a-67769565f895 req-0f1b79d8-7122-49a1-94c3-1c8af279294e service nova] Acquiring lock "refresh_cache-9dd8b658-886a-4f11-ab9b-059c4265436c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.518928] env[63355]: DEBUG oslo_concurrency.lockutils [req-3f3f94de-fb45-4e56-b30a-67769565f895 req-0f1b79d8-7122-49a1-94c3-1c8af279294e service nova] Acquired lock "refresh_cache-9dd8b658-886a-4f11-ab9b-059c4265436c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.519123] env[63355]: DEBUG nova.network.neutron [req-3f3f94de-fb45-4e56-b30a-67769565f895 req-0f1b79d8-7122-49a1-94c3-1c8af279294e service nova] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Refreshing network info cache for port a3482ef0-da11-44f3-84af-7490a7523ddd {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 750.670672] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.670819] env[63355]: ERROR nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e9eae179-ea5b-4a52-8f83-b85e299d41d3, please check neutron logs for more information. [ 750.670819] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Traceback (most recent call last): [ 750.670819] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.670819] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] self.driver.spawn(context, instance, image_meta, [ 750.670819] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 750.670819] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.670819] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.670819] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] vm_ref = self.build_virtual_machine(instance, [ 750.670819] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.670819] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.670819] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] for vif in network_info: [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] return self._sync_wrapper(fn, *args, **kwargs) [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] self.wait() [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] self[:] = self._gt.wait() [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] return self._exit_event.wait() [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] result = hub.switch() [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 750.671127] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] return self.greenlet.switch() [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] result = function(*args, **kwargs) [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] return func(*args, **kwargs) [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] raise e [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] nwinfo = self.network_api.allocate_for_instance( [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] created_port_ids = self._update_ports_for_instance( [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] with excutils.save_and_reraise_exception(): [ 750.671455] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] self.force_reraise() [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] raise self.value [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] updated_port = self._update_port( [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] _ensure_no_port_binding_failure(port) [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] raise exception.PortBindingFailed(port_id=port['id']) [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] nova.exception.PortBindingFailed: Binding failed for port e9eae179-ea5b-4a52-8f83-b85e299d41d3, please check neutron logs for more information. [ 750.672060] env[63355]: ERROR nova.compute.manager [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] [ 750.672384] env[63355]: DEBUG nova.compute.utils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Binding failed for port e9eae179-ea5b-4a52-8f83-b85e299d41d3, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 750.672804] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.959s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.674865] env[63355]: INFO nova.compute.claims [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.678405] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Build of instance 12f9b2b4-3c78-43c2-bb9b-d44b81c83123 was re-scheduled: Binding failed for port e9eae179-ea5b-4a52-8f83-b85e299d41d3, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 750.678754] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 750.679035] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.679183] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquired lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.679343] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.694924] env[63355]: DEBUG nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.722070] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.722363] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.722542] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.722723] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.722866] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.723015] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.723231] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.723388] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.723554] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.723710] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.723880] env[63355]: DEBUG nova.virt.hardware [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.725476] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd89d84-c7e8-465e-a4f4-da085e7232ab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.736082] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6abe95-b1e5-4c7a-8a50-3ed0e197e383 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.772051] env[63355]: ERROR nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a3482ef0-da11-44f3-84af-7490a7523ddd, please check neutron logs for more information. [ 750.772051] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 750.772051] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.772051] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 750.772051] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.772051] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 750.772051] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.772051] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 750.772051] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.772051] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 750.772051] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.772051] env[63355]: ERROR nova.compute.manager raise self.value [ 750.772051] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.772051] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 750.772051] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.772051] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 750.772635] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.772635] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 750.772635] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a3482ef0-da11-44f3-84af-7490a7523ddd, please check neutron logs for more information. [ 750.772635] env[63355]: ERROR nova.compute.manager [ 750.775141] env[63355]: Traceback (most recent call last): [ 750.775141] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 750.775141] env[63355]: listener.cb(fileno) [ 750.775141] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.775141] env[63355]: result = function(*args, **kwargs) [ 750.775141] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.775141] env[63355]: return func(*args, **kwargs) [ 750.775141] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.775141] env[63355]: raise e [ 750.775141] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.775141] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 750.775141] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.775141] env[63355]: created_port_ids = self._update_ports_for_instance( [ 750.775141] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.775141] env[63355]: with excutils.save_and_reraise_exception(): [ 750.775141] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.775141] env[63355]: self.force_reraise() [ 750.775141] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.775141] env[63355]: raise self.value [ 750.775141] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.775141] env[63355]: updated_port = self._update_port( [ 750.775141] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.775141] env[63355]: _ensure_no_port_binding_failure(port) [ 750.775141] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.775141] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 750.775141] env[63355]: nova.exception.PortBindingFailed: Binding failed for port a3482ef0-da11-44f3-84af-7490a7523ddd, please check neutron logs for more information. [ 750.775141] env[63355]: Removing descriptor: 18 [ 750.776025] env[63355]: ERROR nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a3482ef0-da11-44f3-84af-7490a7523ddd, please check neutron logs for more information. [ 750.776025] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Traceback (most recent call last): [ 750.776025] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 750.776025] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] yield resources [ 750.776025] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.776025] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] self.driver.spawn(context, instance, image_meta, [ 750.776025] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 750.776025] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.776025] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.776025] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] vm_ref = self.build_virtual_machine(instance, [ 750.776025] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] for vif in network_info: [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] return self._sync_wrapper(fn, *args, **kwargs) [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] self.wait() [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] self[:] = self._gt.wait() [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] return self._exit_event.wait() [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 750.776353] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] result = hub.switch() [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] return self.greenlet.switch() [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] result = function(*args, **kwargs) [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] return func(*args, **kwargs) [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] raise e [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] nwinfo = self.network_api.allocate_for_instance( [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] created_port_ids = self._update_ports_for_instance( [ 750.776709] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] with excutils.save_and_reraise_exception(): [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] self.force_reraise() [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] raise self.value [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] updated_port = self._update_port( [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] _ensure_no_port_binding_failure(port) [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] raise exception.PortBindingFailed(port_id=port['id']) [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] nova.exception.PortBindingFailed: Binding failed for port a3482ef0-da11-44f3-84af-7490a7523ddd, please check neutron logs for more information. [ 750.777101] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] [ 750.777476] env[63355]: INFO nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Terminating instance [ 750.778459] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "refresh_cache-9dd8b658-886a-4f11-ab9b-059c4265436c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.038642] env[63355]: DEBUG nova.network.neutron [req-3f3f94de-fb45-4e56-b30a-67769565f895 req-0f1b79d8-7122-49a1-94c3-1c8af279294e service nova] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.159769] env[63355]: DEBUG nova.network.neutron [req-3f3f94de-fb45-4e56-b30a-67769565f895 req-0f1b79d8-7122-49a1-94c3-1c8af279294e service nova] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.218156] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.330058] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.664024] env[63355]: DEBUG oslo_concurrency.lockutils [req-3f3f94de-fb45-4e56-b30a-67769565f895 req-0f1b79d8-7122-49a1-94c3-1c8af279294e service nova] Releasing lock "refresh_cache-9dd8b658-886a-4f11-ab9b-059c4265436c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.664024] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquired lock "refresh_cache-9dd8b658-886a-4f11-ab9b-059c4265436c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.664271] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.832717] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Releasing lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.832889] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 751.832979] env[63355]: DEBUG nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.833164] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.850246] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.055968] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b375eb20-7880-4fdb-9900-e067a342c528 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.063564] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70dd9b5-bb89-442c-9bf6-b4eb345596a4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.093696] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b99ff2-e161-4eb1-b70a-870ae392eca7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.100336] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a049a2-4437-42eb-96b7-4ff67111bba4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.113184] env[63355]: DEBUG nova.compute.provider_tree [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.181251] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.303176] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.353231] env[63355]: DEBUG nova.network.neutron [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.592340] env[63355]: DEBUG nova.compute.manager [req-66c9fac4-4e79-4e50-b4c5-89d1d4a23ac3 req-da82dcdf-b821-44f5-bad6-c8c2fac6ff52 service nova] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Received event network-vif-deleted-a3482ef0-da11-44f3-84af-7490a7523ddd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.616684] env[63355]: DEBUG nova.scheduler.client.report [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.806065] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Releasing lock "refresh_cache-9dd8b658-886a-4f11-ab9b-059c4265436c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.806285] env[63355]: DEBUG nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.806481] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 752.806816] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-777d640c-6f7f-48b5-85be-15150c7dcfed {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.816103] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36550320-6999-4ddb-b74f-8b5472394593 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.837443] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9dd8b658-886a-4f11-ab9b-059c4265436c could not be found. [ 752.837673] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.837851] env[63355]: INFO nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 752.838246] env[63355]: DEBUG oslo.service.loopingcall [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.838476] env[63355]: DEBUG nova.compute.manager [-] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.838571] env[63355]: DEBUG nova.network.neutron [-] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.856372] env[63355]: INFO nova.compute.manager [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Took 1.02 seconds to deallocate network for instance. [ 752.859522] env[63355]: DEBUG nova.network.neutron [-] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.121556] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.122079] env[63355]: DEBUG nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 753.124782] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.333s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.126189] env[63355]: INFO nova.compute.claims [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.365977] env[63355]: DEBUG nova.network.neutron [-] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.632936] env[63355]: DEBUG nova.compute.utils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 753.634220] env[63355]: DEBUG nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 753.634413] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 753.682798] env[63355]: DEBUG nova.policy [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1c5de9a1f739467bbbf4b7e6667fdd43', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1bed1ab2806a4ebc8438aa66484683ab', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 753.870213] env[63355]: INFO nova.compute.manager [-] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Took 1.03 seconds to deallocate network for instance. [ 753.873070] env[63355]: DEBUG nova.compute.claims [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 753.873354] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.890668] env[63355]: INFO nova.scheduler.client.report [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Deleted allocations for instance 12f9b2b4-3c78-43c2-bb9b-d44b81c83123 [ 754.057179] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Successfully created port: a5cfc749-0746-45b0-b6b9-c9f92637b2d1 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.138336] env[63355]: DEBUG nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 754.401073] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26018435-eb5b-41f6-9d40-8748e2c4084d tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "12f9b2b4-3c78-43c2-bb9b-d44b81c83123" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.279s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.402246] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "12f9b2b4-3c78-43c2-bb9b-d44b81c83123" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 34.848s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.402527] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "12f9b2b4-3c78-43c2-bb9b-d44b81c83123-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.402743] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "12f9b2b4-3c78-43c2-bb9b-d44b81c83123-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.402913] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "12f9b2b4-3c78-43c2-bb9b-d44b81c83123-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.410966] env[63355]: INFO nova.compute.manager [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Terminating instance [ 754.412599] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquiring lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.412754] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Acquired lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.412927] env[63355]: DEBUG nova.network.neutron [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.520369] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10e06e8-4b8b-482c-b362-beb0a647fd33 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.527827] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c78d4d5-f023-4a54-93b3-eef31720ab33 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.563602] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc1aed4-aaf2-4be3-af9c-5519486bb993 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.570986] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc48c43-6177-4926-913e-dc4a7197728e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.589361] env[63355]: DEBUG nova.compute.provider_tree [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.911786] env[63355]: DEBUG nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 754.935660] env[63355]: DEBUG nova.network.neutron [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.024127] env[63355]: DEBUG nova.network.neutron [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.091931] env[63355]: DEBUG nova.scheduler.client.report [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.148813] env[63355]: DEBUG nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 755.179979] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.180218] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.180376] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.180550] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.180689] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.180927] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.181065] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.181238] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.181400] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.181553] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.181718] env[63355]: DEBUG nova.virt.hardware [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.182575] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44fd1e6-88a5-41fb-9987-ffc3472195f2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.191322] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6c6a34-35b7-4e5a-8f6e-7eb90a5cb2f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.309628] env[63355]: DEBUG nova.compute.manager [req-1fbc65f1-0864-4256-827d-36bc7bdd22f4 req-60dbb93e-a0d8-4ffe-b19e-2bdaed2c051f service nova] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Received event network-changed-a5cfc749-0746-45b0-b6b9-c9f92637b2d1 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.309825] env[63355]: DEBUG nova.compute.manager [req-1fbc65f1-0864-4256-827d-36bc7bdd22f4 req-60dbb93e-a0d8-4ffe-b19e-2bdaed2c051f service nova] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Refreshing instance network info cache due to event network-changed-a5cfc749-0746-45b0-b6b9-c9f92637b2d1. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 755.310048] env[63355]: DEBUG oslo_concurrency.lockutils [req-1fbc65f1-0864-4256-827d-36bc7bdd22f4 req-60dbb93e-a0d8-4ffe-b19e-2bdaed2c051f service nova] Acquiring lock "refresh_cache-d7589681-fe6c-4dc6-b533-ade8ae634d58" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.310195] env[63355]: DEBUG oslo_concurrency.lockutils [req-1fbc65f1-0864-4256-827d-36bc7bdd22f4 req-60dbb93e-a0d8-4ffe-b19e-2bdaed2c051f service nova] Acquired lock "refresh_cache-d7589681-fe6c-4dc6-b533-ade8ae634d58" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.310358] env[63355]: DEBUG nova.network.neutron [req-1fbc65f1-0864-4256-827d-36bc7bdd22f4 req-60dbb93e-a0d8-4ffe-b19e-2bdaed2c051f service nova] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Refreshing network info cache for port a5cfc749-0746-45b0-b6b9-c9f92637b2d1 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 755.435848] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.526813] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Releasing lock "refresh_cache-12f9b2b4-3c78-43c2-bb9b-d44b81c83123" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.527246] env[63355]: DEBUG nova.compute.manager [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 755.527454] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.528125] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c642d0a-dd81-40c8-8c59-229ad5e5d713 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.538744] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66edc4b3-039a-4e42-acfa-4a36c378991f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.567896] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 12f9b2b4-3c78-43c2-bb9b-d44b81c83123 could not be found. [ 755.568113] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 755.568303] env[63355]: INFO nova.compute.manager [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Took 0.04 seconds to destroy the instance on the hypervisor. [ 755.568574] env[63355]: DEBUG oslo.service.loopingcall [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.568775] env[63355]: DEBUG nova.compute.manager [-] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 755.568869] env[63355]: DEBUG nova.network.neutron [-] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 755.586034] env[63355]: DEBUG nova.network.neutron [-] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.597952] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.598542] env[63355]: DEBUG nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 755.601118] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 33.203s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.612134] env[63355]: ERROR nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a5cfc749-0746-45b0-b6b9-c9f92637b2d1, please check neutron logs for more information. [ 755.612134] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 755.612134] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.612134] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 755.612134] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.612134] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 755.612134] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.612134] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 755.612134] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.612134] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 755.612134] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.612134] env[63355]: ERROR nova.compute.manager raise self.value [ 755.612134] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.612134] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 755.612134] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.612134] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 755.612616] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.612616] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 755.612616] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a5cfc749-0746-45b0-b6b9-c9f92637b2d1, please check neutron logs for more information. [ 755.612616] env[63355]: ERROR nova.compute.manager [ 755.612616] env[63355]: Traceback (most recent call last): [ 755.612616] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 755.612616] env[63355]: listener.cb(fileno) [ 755.612616] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.612616] env[63355]: result = function(*args, **kwargs) [ 755.612616] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.612616] env[63355]: return func(*args, **kwargs) [ 755.612616] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.612616] env[63355]: raise e [ 755.612616] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.612616] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 755.612616] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.612616] env[63355]: created_port_ids = self._update_ports_for_instance( [ 755.612616] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.612616] env[63355]: with excutils.save_and_reraise_exception(): [ 755.612616] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.612616] env[63355]: self.force_reraise() [ 755.612616] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.612616] env[63355]: raise self.value [ 755.612616] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.612616] env[63355]: updated_port = self._update_port( [ 755.612616] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.612616] env[63355]: _ensure_no_port_binding_failure(port) [ 755.612616] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.612616] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 755.613658] env[63355]: nova.exception.PortBindingFailed: Binding failed for port a5cfc749-0746-45b0-b6b9-c9f92637b2d1, please check neutron logs for more information. [ 755.613658] env[63355]: Removing descriptor: 18 [ 755.613658] env[63355]: ERROR nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a5cfc749-0746-45b0-b6b9-c9f92637b2d1, please check neutron logs for more information. [ 755.613658] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Traceback (most recent call last): [ 755.613658] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 755.613658] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] yield resources [ 755.613658] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 755.613658] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] self.driver.spawn(context, instance, image_meta, [ 755.613658] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 755.613658] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.613658] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.613658] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] vm_ref = self.build_virtual_machine(instance, [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] for vif in network_info: [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] return self._sync_wrapper(fn, *args, **kwargs) [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] self.wait() [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] self[:] = self._gt.wait() [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] return self._exit_event.wait() [ 755.614032] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] result = hub.switch() [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] return self.greenlet.switch() [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] result = function(*args, **kwargs) [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] return func(*args, **kwargs) [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] raise e [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] nwinfo = self.network_api.allocate_for_instance( [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.615378] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] created_port_ids = self._update_ports_for_instance( [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] with excutils.save_and_reraise_exception(): [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] self.force_reraise() [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] raise self.value [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] updated_port = self._update_port( [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] _ensure_no_port_binding_failure(port) [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.615816] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] raise exception.PortBindingFailed(port_id=port['id']) [ 755.620630] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] nova.exception.PortBindingFailed: Binding failed for port a5cfc749-0746-45b0-b6b9-c9f92637b2d1, please check neutron logs for more information. [ 755.620630] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] [ 755.620630] env[63355]: INFO nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Terminating instance [ 755.620630] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Acquiring lock "refresh_cache-d7589681-fe6c-4dc6-b533-ade8ae634d58" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.840437] env[63355]: DEBUG nova.network.neutron [req-1fbc65f1-0864-4256-827d-36bc7bdd22f4 req-60dbb93e-a0d8-4ffe-b19e-2bdaed2c051f service nova] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.042833] env[63355]: DEBUG nova.network.neutron [req-1fbc65f1-0864-4256-827d-36bc7bdd22f4 req-60dbb93e-a0d8-4ffe-b19e-2bdaed2c051f service nova] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.087840] env[63355]: DEBUG nova.network.neutron [-] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.106158] env[63355]: DEBUG nova.compute.utils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.111807] env[63355]: DEBUG nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 756.111807] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 756.197296] env[63355]: DEBUG nova.policy [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bea2c2f3fb144cc9adba6ffda59627d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2bcc167f94446c88462022a4a39380e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.545885] env[63355]: DEBUG oslo_concurrency.lockutils [req-1fbc65f1-0864-4256-827d-36bc7bdd22f4 req-60dbb93e-a0d8-4ffe-b19e-2bdaed2c051f service nova] Releasing lock "refresh_cache-d7589681-fe6c-4dc6-b533-ade8ae634d58" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.548398] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Acquired lock "refresh_cache-d7589681-fe6c-4dc6-b533-ade8ae634d58" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.548589] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 756.571413] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Successfully created port: 0f896ea1-ac3c-418a-9a64-5efa608ee2bd {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.589979] env[63355]: INFO nova.compute.manager [-] [instance: 12f9b2b4-3c78-43c2-bb9b-d44b81c83123] Took 1.02 seconds to deallocate network for instance. [ 756.592973] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5932ab2-60fd-4ce7-bf66-d7ae84f434f4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.603464] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dba9d39-33d8-4ffe-820e-7d8b175d5e0e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.633654] env[63355]: DEBUG nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 756.637803] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63945e68-190e-4e57-906d-4a8b6d8e2664 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.647398] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2be5c8-034c-4861-b0fc-2861fa670b55 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.662303] env[63355]: DEBUG nova.compute.provider_tree [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.072701] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.164812] env[63355]: DEBUG nova.scheduler.client.report [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.205941] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.545226] env[63355]: DEBUG nova.compute.manager [req-1a9afcec-a44f-48fd-8d17-7036069ca1a1 req-2598d6b2-4227-4c57-bc87-ec0d62bd73e7 service nova] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Received event network-vif-deleted-a5cfc749-0746-45b0-b6b9-c9f92637b2d1 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.628904] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b678627b-901a-4f32-8b87-49c1d3fda650 tempest-ListServersNegativeTestJSON-720562887 tempest-ListServersNegativeTestJSON-720562887-project-member] Lock "12f9b2b4-3c78-43c2-bb9b-d44b81c83123" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.226s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.645723] env[63355]: DEBUG nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 757.670399] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.670979] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.671559] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.673274] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.673274] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.673274] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.673274] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.673274] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.673532] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.673673] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.673996] env[63355]: DEBUG nova.virt.hardware [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.675321] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.074s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.675448] env[63355]: ERROR nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 47293971-a17c-4da3-a20d-458ea560a914, please check neutron logs for more information. [ 757.675448] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Traceback (most recent call last): [ 757.675448] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.675448] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] self.driver.spawn(context, instance, image_meta, [ 757.675448] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 757.675448] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.675448] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.675448] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] vm_ref = self.build_virtual_machine(instance, [ 757.675448] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.675448] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.675448] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] for vif in network_info: [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] return self._sync_wrapper(fn, *args, **kwargs) [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] self.wait() [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] self[:] = self._gt.wait() [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] return self._exit_event.wait() [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] current.throw(*self._exc) [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.675755] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] result = function(*args, **kwargs) [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] return func(*args, **kwargs) [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] raise e [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] nwinfo = self.network_api.allocate_for_instance( [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] created_port_ids = self._update_ports_for_instance( [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] with excutils.save_and_reraise_exception(): [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] self.force_reraise() [ 757.676152] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.676501] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] raise self.value [ 757.676501] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.676501] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] updated_port = self._update_port( [ 757.676501] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.676501] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] _ensure_no_port_binding_failure(port) [ 757.676501] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.676501] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] raise exception.PortBindingFailed(port_id=port['id']) [ 757.676501] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] nova.exception.PortBindingFailed: Binding failed for port 47293971-a17c-4da3-a20d-458ea560a914, please check neutron logs for more information. [ 757.676501] env[63355]: ERROR nova.compute.manager [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] [ 757.676813] env[63355]: DEBUG nova.compute.utils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Binding failed for port 47293971-a17c-4da3-a20d-458ea560a914, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 757.678565] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71c922c-2a34-45cd-9b83-665a27eb6498 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.681815] env[63355]: DEBUG nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Build of instance ad05eda3-efd7-4040-be9c-6a47df02889b was re-scheduled: Binding failed for port 47293971-a17c-4da3-a20d-458ea560a914, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 757.682338] env[63355]: DEBUG nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 757.682636] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquiring lock "refresh_cache-ad05eda3-efd7-4040-be9c-6a47df02889b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.682840] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Acquired lock "refresh_cache-ad05eda3-efd7-4040-be9c-6a47df02889b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.683058] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.684314] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 33.390s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.693404] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6216ed-954a-450e-9ead-9d6435ea3ec2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.710299] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Releasing lock "refresh_cache-d7589681-fe6c-4dc6-b533-ade8ae634d58" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.710768] env[63355]: DEBUG nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 757.711029] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.711374] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42138a7d-db65-4b6d-9f7e-ce945f254300 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.720297] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22debb6-b279-4f86-9ae6-788e47949e0c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.742232] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7589681-fe6c-4dc6-b533-ade8ae634d58 could not be found. [ 757.742543] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 757.742798] env[63355]: INFO nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Took 0.03 seconds to destroy the instance on the hypervisor. [ 757.743098] env[63355]: DEBUG oslo.service.loopingcall [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.743424] env[63355]: DEBUG nova.compute.manager [-] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.743501] env[63355]: DEBUG nova.network.neutron [-] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 757.760614] env[63355]: DEBUG nova.network.neutron [-] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.029752] env[63355]: ERROR nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0f896ea1-ac3c-418a-9a64-5efa608ee2bd, please check neutron logs for more information. [ 758.029752] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 758.029752] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.029752] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 758.029752] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.029752] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 758.029752] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.029752] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 758.029752] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.029752] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 758.029752] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.029752] env[63355]: ERROR nova.compute.manager raise self.value [ 758.029752] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.029752] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 758.029752] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.029752] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 758.030225] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.030225] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 758.030225] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0f896ea1-ac3c-418a-9a64-5efa608ee2bd, please check neutron logs for more information. [ 758.030225] env[63355]: ERROR nova.compute.manager [ 758.030225] env[63355]: Traceback (most recent call last): [ 758.030225] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 758.030225] env[63355]: listener.cb(fileno) [ 758.030225] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.030225] env[63355]: result = function(*args, **kwargs) [ 758.030225] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.030225] env[63355]: return func(*args, **kwargs) [ 758.030225] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.030225] env[63355]: raise e [ 758.030225] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.030225] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 758.030225] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.030225] env[63355]: created_port_ids = self._update_ports_for_instance( [ 758.030225] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.030225] env[63355]: with excutils.save_and_reraise_exception(): [ 758.030225] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.030225] env[63355]: self.force_reraise() [ 758.030225] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.030225] env[63355]: raise self.value [ 758.030225] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.030225] env[63355]: updated_port = self._update_port( [ 758.030225] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.030225] env[63355]: _ensure_no_port_binding_failure(port) [ 758.030225] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.030225] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 758.031136] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 0f896ea1-ac3c-418a-9a64-5efa608ee2bd, please check neutron logs for more information. [ 758.031136] env[63355]: Removing descriptor: 18 [ 758.031136] env[63355]: ERROR nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0f896ea1-ac3c-418a-9a64-5efa608ee2bd, please check neutron logs for more information. [ 758.031136] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Traceback (most recent call last): [ 758.031136] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 758.031136] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] yield resources [ 758.031136] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.031136] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] self.driver.spawn(context, instance, image_meta, [ 758.031136] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 758.031136] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.031136] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.031136] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] vm_ref = self.build_virtual_machine(instance, [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] for vif in network_info: [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] return self._sync_wrapper(fn, *args, **kwargs) [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] self.wait() [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] self[:] = self._gt.wait() [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] return self._exit_event.wait() [ 758.031509] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] result = hub.switch() [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] return self.greenlet.switch() [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] result = function(*args, **kwargs) [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] return func(*args, **kwargs) [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] raise e [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] nwinfo = self.network_api.allocate_for_instance( [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.031847] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] created_port_ids = self._update_ports_for_instance( [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] with excutils.save_and_reraise_exception(): [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] self.force_reraise() [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] raise self.value [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] updated_port = self._update_port( [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] _ensure_no_port_binding_failure(port) [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.032191] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] raise exception.PortBindingFailed(port_id=port['id']) [ 758.033412] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] nova.exception.PortBindingFailed: Binding failed for port 0f896ea1-ac3c-418a-9a64-5efa608ee2bd, please check neutron logs for more information. [ 758.033412] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] [ 758.033412] env[63355]: INFO nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Terminating instance [ 758.033412] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "refresh_cache-17a904fe-ab4b-46ed-98c8-175987710643" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.033541] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquired lock "refresh_cache-17a904fe-ab4b-46ed-98c8-175987710643" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.033689] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.209691] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.264360] env[63355]: DEBUG nova.network.neutron [-] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.273768] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.556026] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.592981] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0506a5d-873d-428e-ba5c-2286346d4711 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.601884] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cc186d-d5fd-4e2e-be07-60a1d57a1a00 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.631195] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d206a75e-5860-4a4b-a7eb-edb153a76693 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.640315] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47bfd91-5d88-4648-a7bc-e3ec7b7f3672 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.653319] env[63355]: DEBUG nova.compute.provider_tree [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.708532] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.767520] env[63355]: INFO nova.compute.manager [-] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Took 1.02 seconds to deallocate network for instance. [ 758.770085] env[63355]: DEBUG nova.compute.claims [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 758.770085] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.776743] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Releasing lock "refresh_cache-ad05eda3-efd7-4040-be9c-6a47df02889b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.776743] env[63355]: DEBUG nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 758.776866] env[63355]: DEBUG nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.777481] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.796157] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.157700] env[63355]: DEBUG nova.scheduler.client.report [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.210714] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Releasing lock "refresh_cache-17a904fe-ab4b-46ed-98c8-175987710643" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.211162] env[63355]: DEBUG nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.211356] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.211643] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-246688f6-d167-494f-b737-99efc57ae433 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.220877] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21587b38-c35c-4ee0-9cb4-471720a2f299 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.242272] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 17a904fe-ab4b-46ed-98c8-175987710643 could not be found. [ 759.242493] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.242674] env[63355]: INFO nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Took 0.03 seconds to destroy the instance on the hypervisor. [ 759.242912] env[63355]: DEBUG oslo.service.loopingcall [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.243148] env[63355]: DEBUG nova.compute.manager [-] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.243265] env[63355]: DEBUG nova.network.neutron [-] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.267210] env[63355]: DEBUG nova.network.neutron [-] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.300611] env[63355]: DEBUG nova.network.neutron [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.575433] env[63355]: DEBUG nova.compute.manager [req-ea3d9f62-e664-4175-9b45-cc02bfc92348 req-07530b47-538a-4ed4-8e64-8ee6175a2f37 service nova] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Received event network-changed-0f896ea1-ac3c-418a-9a64-5efa608ee2bd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.575650] env[63355]: DEBUG nova.compute.manager [req-ea3d9f62-e664-4175-9b45-cc02bfc92348 req-07530b47-538a-4ed4-8e64-8ee6175a2f37 service nova] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Refreshing instance network info cache due to event network-changed-0f896ea1-ac3c-418a-9a64-5efa608ee2bd. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.575860] env[63355]: DEBUG oslo_concurrency.lockutils [req-ea3d9f62-e664-4175-9b45-cc02bfc92348 req-07530b47-538a-4ed4-8e64-8ee6175a2f37 service nova] Acquiring lock "refresh_cache-17a904fe-ab4b-46ed-98c8-175987710643" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.576011] env[63355]: DEBUG oslo_concurrency.lockutils [req-ea3d9f62-e664-4175-9b45-cc02bfc92348 req-07530b47-538a-4ed4-8e64-8ee6175a2f37 service nova] Acquired lock "refresh_cache-17a904fe-ab4b-46ed-98c8-175987710643" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.576181] env[63355]: DEBUG nova.network.neutron [req-ea3d9f62-e664-4175-9b45-cc02bfc92348 req-07530b47-538a-4ed4-8e64-8ee6175a2f37 service nova] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Refreshing network info cache for port 0f896ea1-ac3c-418a-9a64-5efa608ee2bd {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 759.662504] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.663154] env[63355]: ERROR nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6973d5f6-e86d-4734-85ff-ff458f776431, please check neutron logs for more information. [ 759.663154] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Traceback (most recent call last): [ 759.663154] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.663154] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] self.driver.spawn(context, instance, image_meta, [ 759.663154] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 759.663154] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.663154] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.663154] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] vm_ref = self.build_virtual_machine(instance, [ 759.663154] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.663154] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.663154] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] for vif in network_info: [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] return self._sync_wrapper(fn, *args, **kwargs) [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] self.wait() [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] self[:] = self._gt.wait() [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] return self._exit_event.wait() [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] current.throw(*self._exc) [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.663653] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] result = function(*args, **kwargs) [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] return func(*args, **kwargs) [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] raise e [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] nwinfo = self.network_api.allocate_for_instance( [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] created_port_ids = self._update_ports_for_instance( [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] with excutils.save_and_reraise_exception(): [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] self.force_reraise() [ 759.664053] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.664485] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] raise self.value [ 759.664485] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.664485] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] updated_port = self._update_port( [ 759.664485] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.664485] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] _ensure_no_port_binding_failure(port) [ 759.664485] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.664485] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] raise exception.PortBindingFailed(port_id=port['id']) [ 759.664485] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] nova.exception.PortBindingFailed: Binding failed for port 6973d5f6-e86d-4734-85ff-ff458f776431, please check neutron logs for more information. [ 759.664485] env[63355]: ERROR nova.compute.manager [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] [ 759.664485] env[63355]: DEBUG nova.compute.utils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Binding failed for port 6973d5f6-e86d-4734-85ff-ff458f776431, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 759.665082] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.755s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.667927] env[63355]: DEBUG nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Build of instance 9478ce7a-99c6-4e95-82b0-f8a71ce94a90 was re-scheduled: Binding failed for port 6973d5f6-e86d-4734-85ff-ff458f776431, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 759.668359] env[63355]: DEBUG nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 759.668657] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Acquiring lock "refresh_cache-9478ce7a-99c6-4e95-82b0-f8a71ce94a90" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.668721] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Acquired lock "refresh_cache-9478ce7a-99c6-4e95-82b0-f8a71ce94a90" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.668913] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.769470] env[63355]: DEBUG nova.network.neutron [-] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.803126] env[63355]: INFO nova.compute.manager [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] [instance: ad05eda3-efd7-4040-be9c-6a47df02889b] Took 1.03 seconds to deallocate network for instance. [ 760.096278] env[63355]: DEBUG nova.network.neutron [req-ea3d9f62-e664-4175-9b45-cc02bfc92348 req-07530b47-538a-4ed4-8e64-8ee6175a2f37 service nova] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.188028] env[63355]: DEBUG nova.network.neutron [req-ea3d9f62-e664-4175-9b45-cc02bfc92348 req-07530b47-538a-4ed4-8e64-8ee6175a2f37 service nova] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.189401] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.235267] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.272156] env[63355]: INFO nova.compute.manager [-] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Took 1.03 seconds to deallocate network for instance. [ 760.274346] env[63355]: DEBUG nova.compute.claims [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 760.274585] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.531325] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb2469a-f266-4745-9dd3-19a8c621b63a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.538923] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080c8cb8-1fde-4be2-a672-3a8a42276f17 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.569479] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f411d65e-5909-4e62-8ec4-59253893e1af {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.576627] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d40dfe-9554-4a48-b666-86334e0f51ab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.589285] env[63355]: DEBUG nova.compute.provider_tree [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.692372] env[63355]: DEBUG oslo_concurrency.lockutils [req-ea3d9f62-e664-4175-9b45-cc02bfc92348 req-07530b47-538a-4ed4-8e64-8ee6175a2f37 service nova] Releasing lock "refresh_cache-17a904fe-ab4b-46ed-98c8-175987710643" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.692665] env[63355]: DEBUG nova.compute.manager [req-ea3d9f62-e664-4175-9b45-cc02bfc92348 req-07530b47-538a-4ed4-8e64-8ee6175a2f37 service nova] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Received event network-vif-deleted-0f896ea1-ac3c-418a-9a64-5efa608ee2bd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.738619] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Releasing lock "refresh_cache-9478ce7a-99c6-4e95-82b0-f8a71ce94a90" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.738829] env[63355]: DEBUG nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 760.739065] env[63355]: DEBUG nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.740039] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.757418] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.839212] env[63355]: INFO nova.scheduler.client.report [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Deleted allocations for instance ad05eda3-efd7-4040-be9c-6a47df02889b [ 761.092105] env[63355]: DEBUG nova.scheduler.client.report [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.266245] env[63355]: DEBUG nova.network.neutron [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.349219] env[63355]: DEBUG oslo_concurrency.lockutils [None req-82c4e961-d1f9-4dfb-982b-a831918a3b08 tempest-MigrationsAdminTest-435433884 tempest-MigrationsAdminTest-435433884-project-member] Lock "ad05eda3-efd7-4040-be9c-6a47df02889b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.772s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.598295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.933s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.599069] env[63355]: ERROR nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ac49af64-6ba9-4980-b8dc-ac660a40adae, please check neutron logs for more information. [ 761.599069] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Traceback (most recent call last): [ 761.599069] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.599069] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] self.driver.spawn(context, instance, image_meta, [ 761.599069] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 761.599069] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.599069] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.599069] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] vm_ref = self.build_virtual_machine(instance, [ 761.599069] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.599069] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.599069] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] for vif in network_info: [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] return self._sync_wrapper(fn, *args, **kwargs) [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] self.wait() [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] self[:] = self._gt.wait() [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] return self._exit_event.wait() [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] current.throw(*self._exc) [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.599621] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] result = function(*args, **kwargs) [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] return func(*args, **kwargs) [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] raise e [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] nwinfo = self.network_api.allocate_for_instance( [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] created_port_ids = self._update_ports_for_instance( [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] with excutils.save_and_reraise_exception(): [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] self.force_reraise() [ 761.600055] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.600449] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] raise self.value [ 761.600449] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.600449] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] updated_port = self._update_port( [ 761.600449] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.600449] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] _ensure_no_port_binding_failure(port) [ 761.600449] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.600449] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] raise exception.PortBindingFailed(port_id=port['id']) [ 761.600449] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] nova.exception.PortBindingFailed: Binding failed for port ac49af64-6ba9-4980-b8dc-ac660a40adae, please check neutron logs for more information. [ 761.600449] env[63355]: ERROR nova.compute.manager [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] [ 761.600449] env[63355]: DEBUG nova.compute.utils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Binding failed for port ac49af64-6ba9-4980-b8dc-ac660a40adae, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.601055] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.248s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.602669] env[63355]: INFO nova.compute.claims [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.605443] env[63355]: DEBUG nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Build of instance b7dcc653-a430-48c4-9185-3e8200cf1005 was re-scheduled: Binding failed for port ac49af64-6ba9-4980-b8dc-ac660a40adae, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 761.605880] env[63355]: DEBUG nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 761.606101] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "refresh_cache-b7dcc653-a430-48c4-9185-3e8200cf1005" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.606251] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "refresh_cache-b7dcc653-a430-48c4-9185-3e8200cf1005" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.606409] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.768367] env[63355]: INFO nova.compute.manager [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] [instance: 9478ce7a-99c6-4e95-82b0-f8a71ce94a90] Took 1.03 seconds to deallocate network for instance. [ 761.851274] env[63355]: DEBUG nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.132311] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.228258] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.375848] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.730809] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "refresh_cache-b7dcc653-a430-48c4-9185-3e8200cf1005" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.731121] env[63355]: DEBUG nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 762.731410] env[63355]: DEBUG nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.731410] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.751463] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.816323] env[63355]: INFO nova.scheduler.client.report [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Deleted allocations for instance 9478ce7a-99c6-4e95-82b0-f8a71ce94a90 [ 763.006727] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771dc4c8-cea2-41f5-9542-92caecaff92c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.014637] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d9ed53-ba5e-46a6-8fc4-0224e7235a01 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.044717] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3077eec2-0408-4bc2-a70d-49ce1fa2ca54 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.052285] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f846f25b-08a2-42d2-9723-8586cf55c170 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.065263] env[63355]: DEBUG nova.compute.provider_tree [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.258678] env[63355]: DEBUG nova.network.neutron [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.324101] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6f79ebf-1bce-4699-ac70-c2afe1f13d60 tempest-ServersAdminNegativeTestJSON-1716821844 tempest-ServersAdminNegativeTestJSON-1716821844-project-member] Lock "9478ce7a-99c6-4e95-82b0-f8a71ce94a90" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.237s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.567977] env[63355]: DEBUG nova.scheduler.client.report [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.760482] env[63355]: INFO nova.compute.manager [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: b7dcc653-a430-48c4-9185-3e8200cf1005] Took 1.03 seconds to deallocate network for instance. [ 763.829548] env[63355]: DEBUG nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 764.072760] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.073210] env[63355]: DEBUG nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.076269] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.053s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.077331] env[63355]: INFO nova.compute.claims [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.363210] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.584137] env[63355]: DEBUG nova.compute.utils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.585535] env[63355]: DEBUG nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.585702] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 764.626110] env[63355]: DEBUG nova.policy [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bea2c2f3fb144cc9adba6ffda59627d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2bcc167f94446c88462022a4a39380e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 764.795701] env[63355]: INFO nova.scheduler.client.report [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleted allocations for instance b7dcc653-a430-48c4-9185-3e8200cf1005 [ 764.886851] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Successfully created port: aa866120-3950-4692-9cd4-afc27bb3a783 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.089967] env[63355]: DEBUG nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.305263] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75a2f830-9f41-412a-9508-0b855b06cf4f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "b7dcc653-a430-48c4-9185-3e8200cf1005" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.181s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.547314] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4cebd1-0c1c-4cf1-8248-78563eee1691 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.557312] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792071c8-f71f-4b73-b2ef-98ab8527638b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.590495] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cff9c7-87ea-4baf-bb6d-fdffcdecf596 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.604170] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b923801e-30d0-45f3-bbe8-03d38b487678 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.619751] env[63355]: DEBUG nova.compute.provider_tree [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.760549] env[63355]: DEBUG nova.compute.manager [req-e837f758-f4fc-46b8-8a5f-ad21891fc056 req-ade7c7a7-5308-499a-bf88-8e31f5aed504 service nova] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Received event network-changed-aa866120-3950-4692-9cd4-afc27bb3a783 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.760731] env[63355]: DEBUG nova.compute.manager [req-e837f758-f4fc-46b8-8a5f-ad21891fc056 req-ade7c7a7-5308-499a-bf88-8e31f5aed504 service nova] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Refreshing instance network info cache due to event network-changed-aa866120-3950-4692-9cd4-afc27bb3a783. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 765.760948] env[63355]: DEBUG oslo_concurrency.lockutils [req-e837f758-f4fc-46b8-8a5f-ad21891fc056 req-ade7c7a7-5308-499a-bf88-8e31f5aed504 service nova] Acquiring lock "refresh_cache-d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.761600] env[63355]: DEBUG oslo_concurrency.lockutils [req-e837f758-f4fc-46b8-8a5f-ad21891fc056 req-ade7c7a7-5308-499a-bf88-8e31f5aed504 service nova] Acquired lock "refresh_cache-d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.761952] env[63355]: DEBUG nova.network.neutron [req-e837f758-f4fc-46b8-8a5f-ad21891fc056 req-ade7c7a7-5308-499a-bf88-8e31f5aed504 service nova] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Refreshing network info cache for port aa866120-3950-4692-9cd4-afc27bb3a783 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 765.808429] env[63355]: DEBUG nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 765.824065] env[63355]: ERROR nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port aa866120-3950-4692-9cd4-afc27bb3a783, please check neutron logs for more information. [ 765.824065] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 765.824065] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.824065] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 765.824065] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.824065] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 765.824065] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.824065] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 765.824065] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.824065] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 765.824065] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.824065] env[63355]: ERROR nova.compute.manager raise self.value [ 765.824065] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.824065] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 765.824065] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.824065] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 765.824976] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.824976] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 765.824976] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port aa866120-3950-4692-9cd4-afc27bb3a783, please check neutron logs for more information. [ 765.824976] env[63355]: ERROR nova.compute.manager [ 765.824976] env[63355]: Traceback (most recent call last): [ 765.824976] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 765.824976] env[63355]: listener.cb(fileno) [ 765.824976] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.824976] env[63355]: result = function(*args, **kwargs) [ 765.824976] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.824976] env[63355]: return func(*args, **kwargs) [ 765.824976] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.824976] env[63355]: raise e [ 765.824976] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.824976] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 765.824976] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.824976] env[63355]: created_port_ids = self._update_ports_for_instance( [ 765.824976] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.824976] env[63355]: with excutils.save_and_reraise_exception(): [ 765.824976] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.824976] env[63355]: self.force_reraise() [ 765.824976] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.824976] env[63355]: raise self.value [ 765.824976] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.824976] env[63355]: updated_port = self._update_port( [ 765.824976] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.824976] env[63355]: _ensure_no_port_binding_failure(port) [ 765.824976] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.824976] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 765.825842] env[63355]: nova.exception.PortBindingFailed: Binding failed for port aa866120-3950-4692-9cd4-afc27bb3a783, please check neutron logs for more information. [ 765.825842] env[63355]: Removing descriptor: 17 [ 766.100756] env[63355]: DEBUG nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.124919] env[63355]: DEBUG nova.scheduler.client.report [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.131576] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.132043] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.132939] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.133441] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.133708] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.134261] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.134611] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.135158] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.135580] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.135866] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.136173] env[63355]: DEBUG nova.virt.hardware [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.137148] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79a1407-7bad-463e-9d12-e227dced8f95 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.147037] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba3846e-2d5d-4341-910a-69fa601499d3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.167519] env[63355]: ERROR nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port aa866120-3950-4692-9cd4-afc27bb3a783, please check neutron logs for more information. [ 766.167519] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Traceback (most recent call last): [ 766.167519] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 766.167519] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] yield resources [ 766.167519] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.167519] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] self.driver.spawn(context, instance, image_meta, [ 766.167519] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 766.167519] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.167519] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.167519] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] vm_ref = self.build_virtual_machine(instance, [ 766.167519] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] for vif in network_info: [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] return self._sync_wrapper(fn, *args, **kwargs) [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] self.wait() [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] self[:] = self._gt.wait() [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] return self._exit_event.wait() [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.167949] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] current.throw(*self._exc) [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] result = function(*args, **kwargs) [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] return func(*args, **kwargs) [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] raise e [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] nwinfo = self.network_api.allocate_for_instance( [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] created_port_ids = self._update_ports_for_instance( [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] with excutils.save_and_reraise_exception(): [ 766.168356] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] self.force_reraise() [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] raise self.value [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] updated_port = self._update_port( [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] _ensure_no_port_binding_failure(port) [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] raise exception.PortBindingFailed(port_id=port['id']) [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] nova.exception.PortBindingFailed: Binding failed for port aa866120-3950-4692-9cd4-afc27bb3a783, please check neutron logs for more information. [ 766.168747] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] [ 766.168747] env[63355]: INFO nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Terminating instance [ 766.170424] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "refresh_cache-d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.284320] env[63355]: DEBUG nova.network.neutron [req-e837f758-f4fc-46b8-8a5f-ad21891fc056 req-ade7c7a7-5308-499a-bf88-8e31f5aed504 service nova] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.349482] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.383329] env[63355]: DEBUG nova.network.neutron [req-e837f758-f4fc-46b8-8a5f-ad21891fc056 req-ade7c7a7-5308-499a-bf88-8e31f5aed504 service nova] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.645020] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.645020] env[63355]: DEBUG nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.651164] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.149s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.651164] env[63355]: INFO nova.compute.claims [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.886428] env[63355]: DEBUG oslo_concurrency.lockutils [req-e837f758-f4fc-46b8-8a5f-ad21891fc056 req-ade7c7a7-5308-499a-bf88-8e31f5aed504 service nova] Releasing lock "refresh_cache-d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.886837] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquired lock "refresh_cache-d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.889489] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.993224] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.993453] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.155291] env[63355]: DEBUG nova.compute.utils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.160306] env[63355]: DEBUG nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 767.160491] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 767.232424] env[63355]: DEBUG nova.policy [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f442c1a185e4d268c79cb2c20f1f000', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cba27209d3044b35ba03482e7b1973b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.413229] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.511928] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.662233] env[63355]: DEBUG nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.720689] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Successfully created port: 407f76f2-db4c-42a7-a5e7-68033196ad71 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.962071] env[63355]: DEBUG nova.compute.manager [req-d1cdeb12-e716-4c30-8942-86d4aae74055 req-c3cb59dd-eeb6-49de-a534-a66e308c934a service nova] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Received event network-vif-deleted-aa866120-3950-4692-9cd4-afc27bb3a783 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.014738] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Releasing lock "refresh_cache-d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.015238] env[63355]: DEBUG nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.015353] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.015655] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89dfe1be-957b-4a3d-ac52-233573fe712e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.028956] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9299c8-ee63-4b40-b86e-0e1c0350d7b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.054154] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6 could not be found. [ 768.054410] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 768.054995] env[63355]: INFO nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 768.055342] env[63355]: DEBUG oslo.service.loopingcall [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.055768] env[63355]: DEBUG nova.compute.manager [-] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.055768] env[63355]: DEBUG nova.network.neutron [-] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.090385] env[63355]: DEBUG nova.network.neutron [-] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.219385] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50301e24-d182-4c4d-b84f-8898e9210d14 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.227807] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad335d9-e219-450b-89dc-c21616570eb2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.264324] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33aa4bd1-eb74-4daa-9cad-b2a7d79ab8e3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.272884] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18fce994-6323-4a9b-ba05-4a3c7a140694 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.288920] env[63355]: DEBUG nova.compute.provider_tree [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.594239] env[63355]: DEBUG nova.network.neutron [-] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.600939] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.601321] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.675072] env[63355]: DEBUG nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 768.708677] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.708906] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.710082] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.710348] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.710546] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.710723] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.710959] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.711167] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.711391] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.711583] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.714484] env[63355]: DEBUG nova.virt.hardware [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.714484] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423b610f-fbfa-439f-84c3-8f35e0e968bd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.721757] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2d8d5a-1c39-4e18-82ca-8b7173bc599f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.792878] env[63355]: DEBUG nova.scheduler.client.report [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.054526] env[63355]: ERROR nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 407f76f2-db4c-42a7-a5e7-68033196ad71, please check neutron logs for more information. [ 769.054526] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 769.054526] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.054526] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 769.054526] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.054526] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 769.054526] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.054526] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 769.054526] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.054526] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 769.054526] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.054526] env[63355]: ERROR nova.compute.manager raise self.value [ 769.054526] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.054526] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 769.054526] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.054526] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 769.055169] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.055169] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 769.055169] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 407f76f2-db4c-42a7-a5e7-68033196ad71, please check neutron logs for more information. [ 769.055169] env[63355]: ERROR nova.compute.manager [ 769.055169] env[63355]: Traceback (most recent call last): [ 769.055169] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 769.055169] env[63355]: listener.cb(fileno) [ 769.055169] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.055169] env[63355]: result = function(*args, **kwargs) [ 769.055169] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.055169] env[63355]: return func(*args, **kwargs) [ 769.055169] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.055169] env[63355]: raise e [ 769.055169] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.055169] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 769.055169] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.055169] env[63355]: created_port_ids = self._update_ports_for_instance( [ 769.055169] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.055169] env[63355]: with excutils.save_and_reraise_exception(): [ 769.055169] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.055169] env[63355]: self.force_reraise() [ 769.055169] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.055169] env[63355]: raise self.value [ 769.055169] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.055169] env[63355]: updated_port = self._update_port( [ 769.055169] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.055169] env[63355]: _ensure_no_port_binding_failure(port) [ 769.055169] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.055169] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 769.055960] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 407f76f2-db4c-42a7-a5e7-68033196ad71, please check neutron logs for more information. [ 769.055960] env[63355]: Removing descriptor: 17 [ 769.055960] env[63355]: ERROR nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 407f76f2-db4c-42a7-a5e7-68033196ad71, please check neutron logs for more information. [ 769.055960] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Traceback (most recent call last): [ 769.055960] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 769.055960] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] yield resources [ 769.055960] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.055960] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] self.driver.spawn(context, instance, image_meta, [ 769.055960] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 769.055960] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.055960] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.055960] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] vm_ref = self.build_virtual_machine(instance, [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] for vif in network_info: [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] return self._sync_wrapper(fn, *args, **kwargs) [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] self.wait() [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] self[:] = self._gt.wait() [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] return self._exit_event.wait() [ 769.056392] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] result = hub.switch() [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] return self.greenlet.switch() [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] result = function(*args, **kwargs) [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] return func(*args, **kwargs) [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] raise e [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] nwinfo = self.network_api.allocate_for_instance( [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.056829] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] created_port_ids = self._update_ports_for_instance( [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] with excutils.save_and_reraise_exception(): [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] self.force_reraise() [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] raise self.value [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] updated_port = self._update_port( [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] _ensure_no_port_binding_failure(port) [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.057199] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] raise exception.PortBindingFailed(port_id=port['id']) [ 769.057522] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] nova.exception.PortBindingFailed: Binding failed for port 407f76f2-db4c-42a7-a5e7-68033196ad71, please check neutron logs for more information. [ 769.057522] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] [ 769.057522] env[63355]: INFO nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Terminating instance [ 769.058917] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Acquiring lock "refresh_cache-8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.058917] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Acquired lock "refresh_cache-8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.059097] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.098334] env[63355]: INFO nova.compute.manager [-] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Took 1.04 seconds to deallocate network for instance. [ 769.100758] env[63355]: DEBUG nova.compute.claims [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 769.101071] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.297620] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.650s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.298236] env[63355]: DEBUG nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 769.300818] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.353s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.300959] env[63355]: DEBUG nova.objects.instance [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63355) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 769.636412] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquiring lock "b6056441-9ee4-484f-a1d2-077546f2c581" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.636808] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "b6056441-9ee4-484f-a1d2-077546f2c581" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.717745] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.791989] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.805594] env[63355]: DEBUG nova.compute.utils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.809572] env[63355]: DEBUG nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 769.809658] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 769.846618] env[63355]: DEBUG nova.policy [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3fd9a173cf6d4f2c8da17df12a9f3779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87a296f0ecd94879ab2336899fc575bf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.988630] env[63355]: DEBUG nova.compute.manager [req-b853a088-c567-45f3-a978-2d995408b529 req-bec39632-2e40-453d-b410-816be6c07a87 service nova] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Received event network-changed-407f76f2-db4c-42a7-a5e7-68033196ad71 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.988630] env[63355]: DEBUG nova.compute.manager [req-b853a088-c567-45f3-a978-2d995408b529 req-bec39632-2e40-453d-b410-816be6c07a87 service nova] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Refreshing instance network info cache due to event network-changed-407f76f2-db4c-42a7-a5e7-68033196ad71. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.988630] env[63355]: DEBUG oslo_concurrency.lockutils [req-b853a088-c567-45f3-a978-2d995408b529 req-bec39632-2e40-453d-b410-816be6c07a87 service nova] Acquiring lock "refresh_cache-8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.122087] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Successfully created port: fb7b1d5f-c8fa-4ae5-8522-75ae880ae067 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.298026] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Releasing lock "refresh_cache-8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.298026] env[63355]: DEBUG nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 770.298026] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 770.298026] env[63355]: DEBUG oslo_concurrency.lockutils [req-b853a088-c567-45f3-a978-2d995408b529 req-bec39632-2e40-453d-b410-816be6c07a87 service nova] Acquired lock "refresh_cache-8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.298026] env[63355]: DEBUG nova.network.neutron [req-b853a088-c567-45f3-a978-2d995408b529 req-bec39632-2e40-453d-b410-816be6c07a87 service nova] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Refreshing network info cache for port 407f76f2-db4c-42a7-a5e7-68033196ad71 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 770.298345] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ccdedb2c-ef72-4280-a64d-70210f9efa7e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.307646] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0344318-5496-448a-a522-271a7d5ebb6a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.320447] env[63355]: DEBUG nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 770.323669] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9da035c9-8bbf-4f6d-9007-b7c4c7166d5d tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.023s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.324971] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.410s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.325244] env[63355]: DEBUG nova.objects.instance [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lazy-loading 'resources' on Instance uuid b2bf4912-6b88-4ece-95c9-e9fd1cf906df {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 770.337800] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6 could not be found. [ 770.338365] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 770.338599] env[63355]: INFO nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 770.338871] env[63355]: DEBUG oslo.service.loopingcall [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.339665] env[63355]: DEBUG nova.compute.manager [-] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.339757] env[63355]: DEBUG nova.network.neutron [-] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.364991] env[63355]: DEBUG nova.network.neutron [-] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.823427] env[63355]: DEBUG nova.network.neutron [req-b853a088-c567-45f3-a978-2d995408b529 req-bec39632-2e40-453d-b410-816be6c07a87 service nova] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.871100] env[63355]: DEBUG nova.network.neutron [-] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.949599] env[63355]: DEBUG nova.network.neutron [req-b853a088-c567-45f3-a978-2d995408b529 req-bec39632-2e40-453d-b410-816be6c07a87 service nova] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.997757] env[63355]: ERROR nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fb7b1d5f-c8fa-4ae5-8522-75ae880ae067, please check neutron logs for more information. [ 770.997757] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 770.997757] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.997757] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 770.997757] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.997757] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 770.997757] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.997757] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 770.997757] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.997757] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 770.997757] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.997757] env[63355]: ERROR nova.compute.manager raise self.value [ 770.997757] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.997757] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 770.997757] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.997757] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 770.998252] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.998252] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 770.998252] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fb7b1d5f-c8fa-4ae5-8522-75ae880ae067, please check neutron logs for more information. [ 770.998252] env[63355]: ERROR nova.compute.manager [ 770.998252] env[63355]: Traceback (most recent call last): [ 770.998252] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 770.998252] env[63355]: listener.cb(fileno) [ 770.998252] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.998252] env[63355]: result = function(*args, **kwargs) [ 770.998252] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 770.998252] env[63355]: return func(*args, **kwargs) [ 770.998252] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.998252] env[63355]: raise e [ 770.998252] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.998252] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 770.998252] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.998252] env[63355]: created_port_ids = self._update_ports_for_instance( [ 770.998252] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.998252] env[63355]: with excutils.save_and_reraise_exception(): [ 770.998252] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.998252] env[63355]: self.force_reraise() [ 770.998252] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.998252] env[63355]: raise self.value [ 770.998252] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.998252] env[63355]: updated_port = self._update_port( [ 770.998252] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.998252] env[63355]: _ensure_no_port_binding_failure(port) [ 770.998252] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.998252] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 770.999298] env[63355]: nova.exception.PortBindingFailed: Binding failed for port fb7b1d5f-c8fa-4ae5-8522-75ae880ae067, please check neutron logs for more information. [ 770.999298] env[63355]: Removing descriptor: 17 [ 771.235645] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec463f6-add5-4e3c-883d-17f2c8040fc3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.243703] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5559c71e-f16e-46c0-9f08-29944b05eb76 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.275270] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa9a31f-1674-4a6a-8d00-e131c2662311 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.282536] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca25f54-aab9-4475-9827-5594939f053c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.295608] env[63355]: DEBUG nova.compute.provider_tree [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.333026] env[63355]: DEBUG nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 771.356956] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.357225] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.357366] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.357546] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.357719] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.357834] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.358048] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.358202] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.358363] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.358550] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.358678] env[63355]: DEBUG nova.virt.hardware [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.359598] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06774ae-f941-4c31-a213-33b30c6b2a48 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.367691] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3c3015-7e01-4705-a5eb-b146ae7ec592 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.373196] env[63355]: INFO nova.compute.manager [-] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Took 1.03 seconds to deallocate network for instance. [ 771.382893] env[63355]: ERROR nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fb7b1d5f-c8fa-4ae5-8522-75ae880ae067, please check neutron logs for more information. [ 771.382893] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Traceback (most recent call last): [ 771.382893] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 771.382893] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] yield resources [ 771.382893] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 771.382893] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] self.driver.spawn(context, instance, image_meta, [ 771.382893] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 771.382893] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 771.382893] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 771.382893] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] vm_ref = self.build_virtual_machine(instance, [ 771.382893] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] for vif in network_info: [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] return self._sync_wrapper(fn, *args, **kwargs) [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] self.wait() [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] self[:] = self._gt.wait() [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] return self._exit_event.wait() [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 771.383318] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] current.throw(*self._exc) [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] result = function(*args, **kwargs) [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] return func(*args, **kwargs) [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] raise e [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] nwinfo = self.network_api.allocate_for_instance( [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] created_port_ids = self._update_ports_for_instance( [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] with excutils.save_and_reraise_exception(): [ 771.383687] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] self.force_reraise() [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] raise self.value [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] updated_port = self._update_port( [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] _ensure_no_port_binding_failure(port) [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] raise exception.PortBindingFailed(port_id=port['id']) [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] nova.exception.PortBindingFailed: Binding failed for port fb7b1d5f-c8fa-4ae5-8522-75ae880ae067, please check neutron logs for more information. [ 771.384117] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] [ 771.384117] env[63355]: INFO nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Terminating instance [ 771.384749] env[63355]: DEBUG nova.compute.claims [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 771.384922] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.385367] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "refresh_cache-1446c7da-415f-43ee-9ff1-2266bd3038e4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.385591] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquired lock "refresh_cache-1446c7da-415f-43ee-9ff1-2266bd3038e4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.385827] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.450746] env[63355]: DEBUG oslo_concurrency.lockutils [req-b853a088-c567-45f3-a978-2d995408b529 req-bec39632-2e40-453d-b410-816be6c07a87 service nova] Releasing lock "refresh_cache-8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.451019] env[63355]: DEBUG nova.compute.manager [req-b853a088-c567-45f3-a978-2d995408b529 req-bec39632-2e40-453d-b410-816be6c07a87 service nova] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Received event network-vif-deleted-407f76f2-db4c-42a7-a5e7-68033196ad71 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.799101] env[63355]: DEBUG nova.scheduler.client.report [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.905326] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.994207] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.050139] env[63355]: DEBUG nova.compute.manager [req-4559c664-3eb0-484f-9e24-8b25a8da56cd req-15676e7f-a4f2-43fc-b4e1-fa46acd364d5 service nova] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Received event network-changed-fb7b1d5f-c8fa-4ae5-8522-75ae880ae067 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.050352] env[63355]: DEBUG nova.compute.manager [req-4559c664-3eb0-484f-9e24-8b25a8da56cd req-15676e7f-a4f2-43fc-b4e1-fa46acd364d5 service nova] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Refreshing instance network info cache due to event network-changed-fb7b1d5f-c8fa-4ae5-8522-75ae880ae067. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 772.050543] env[63355]: DEBUG oslo_concurrency.lockutils [req-4559c664-3eb0-484f-9e24-8b25a8da56cd req-15676e7f-a4f2-43fc-b4e1-fa46acd364d5 service nova] Acquiring lock "refresh_cache-1446c7da-415f-43ee-9ff1-2266bd3038e4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.307973] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.983s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.311188] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.860s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.311446] env[63355]: DEBUG nova.objects.instance [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lazy-loading 'resources' on Instance uuid e5c80c05-0e48-415a-ac91-a53e1c8819f0 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 772.326606] env[63355]: INFO nova.scheduler.client.report [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Deleted allocations for instance b2bf4912-6b88-4ece-95c9-e9fd1cf906df [ 772.497188] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Releasing lock "refresh_cache-1446c7da-415f-43ee-9ff1-2266bd3038e4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.497610] env[63355]: DEBUG nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 772.497810] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 772.498132] env[63355]: DEBUG oslo_concurrency.lockutils [req-4559c664-3eb0-484f-9e24-8b25a8da56cd req-15676e7f-a4f2-43fc-b4e1-fa46acd364d5 service nova] Acquired lock "refresh_cache-1446c7da-415f-43ee-9ff1-2266bd3038e4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.498298] env[63355]: DEBUG nova.network.neutron [req-4559c664-3eb0-484f-9e24-8b25a8da56cd req-15676e7f-a4f2-43fc-b4e1-fa46acd364d5 service nova] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Refreshing network info cache for port fb7b1d5f-c8fa-4ae5-8522-75ae880ae067 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.499337] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8edbdf0-1e2a-47f9-aff6-8294985b4776 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.509100] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a983559-8367-482e-a7c6-469d52a80b1c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.529954] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1446c7da-415f-43ee-9ff1-2266bd3038e4 could not be found. [ 772.530169] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 772.530344] env[63355]: INFO nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 772.530576] env[63355]: DEBUG oslo.service.loopingcall [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.530790] env[63355]: DEBUG nova.compute.manager [-] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.530881] env[63355]: DEBUG nova.network.neutron [-] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 772.545027] env[63355]: DEBUG nova.network.neutron [-] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.834635] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b5dd1a5e-02cd-4b36-94dc-4484e53727e7 tempest-ServerDiagnosticsV248Test-23539034 tempest-ServerDiagnosticsV248Test-23539034-project-member] Lock "b2bf4912-6b88-4ece-95c9-e9fd1cf906df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.640s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.019609] env[63355]: DEBUG nova.network.neutron [req-4559c664-3eb0-484f-9e24-8b25a8da56cd req-15676e7f-a4f2-43fc-b4e1-fa46acd364d5 service nova] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.046945] env[63355]: DEBUG nova.network.neutron [-] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.116098] env[63355]: DEBUG nova.network.neutron [req-4559c664-3eb0-484f-9e24-8b25a8da56cd req-15676e7f-a4f2-43fc-b4e1-fa46acd364d5 service nova] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.228511] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8c9b74-b866-4cf7-a5af-4c2505a0f967 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.237019] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff33145-4664-4632-ba2d-74d01a467ab8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.269962] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d691a5-7da8-42b7-923a-5ddfd52de922 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.277185] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7073d8-ad3d-4b70-b330-4fab6bfad1e7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.290251] env[63355]: DEBUG nova.compute.provider_tree [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.549684] env[63355]: INFO nova.compute.manager [-] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Took 1.02 seconds to deallocate network for instance. [ 773.552121] env[63355]: DEBUG nova.compute.claims [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 773.552500] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.620611] env[63355]: DEBUG oslo_concurrency.lockutils [req-4559c664-3eb0-484f-9e24-8b25a8da56cd req-15676e7f-a4f2-43fc-b4e1-fa46acd364d5 service nova] Releasing lock "refresh_cache-1446c7da-415f-43ee-9ff1-2266bd3038e4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.620884] env[63355]: DEBUG nova.compute.manager [req-4559c664-3eb0-484f-9e24-8b25a8da56cd req-15676e7f-a4f2-43fc-b4e1-fa46acd364d5 service nova] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Received event network-vif-deleted-fb7b1d5f-c8fa-4ae5-8522-75ae880ae067 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.793436] env[63355]: DEBUG nova.scheduler.client.report [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.298585] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.988s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.300830] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.427s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.322722] env[63355]: INFO nova.scheduler.client.report [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Deleted allocations for instance e5c80c05-0e48-415a-ac91-a53e1c8819f0 [ 774.833088] env[63355]: DEBUG oslo_concurrency.lockutils [None req-179252ba-356e-484e-8234-f84b2cba8578 tempest-ServerShowV254Test-1365464896 tempest-ServerShowV254Test-1365464896-project-member] Lock "e5c80c05-0e48-415a-ac91-a53e1c8819f0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.076s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.253575] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67a8386-0168-49de-aa34-9ea0ac1201a2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.261987] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bef9c34-0874-4e59-a558-f0edd01550fc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.291595] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9732ea39-3ad3-4318-9c4b-3c4dc9384f45 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.299135] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10598b1f-f875-45af-ba54-0e22405d63ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.314874] env[63355]: DEBUG nova.compute.provider_tree [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.818084] env[63355]: DEBUG nova.scheduler.client.report [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.325155] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.024s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.327015] env[63355]: ERROR nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a3482ef0-da11-44f3-84af-7490a7523ddd, please check neutron logs for more information. [ 776.327015] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Traceback (most recent call last): [ 776.327015] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.327015] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] self.driver.spawn(context, instance, image_meta, [ 776.327015] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 776.327015] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.327015] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.327015] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] vm_ref = self.build_virtual_machine(instance, [ 776.327015] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.327015] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.327015] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] for vif in network_info: [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] return self._sync_wrapper(fn, *args, **kwargs) [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] self.wait() [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] self[:] = self._gt.wait() [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] return self._exit_event.wait() [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] result = hub.switch() [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 776.327402] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] return self.greenlet.switch() [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] result = function(*args, **kwargs) [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] return func(*args, **kwargs) [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] raise e [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] nwinfo = self.network_api.allocate_for_instance( [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] created_port_ids = self._update_ports_for_instance( [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] with excutils.save_and_reraise_exception(): [ 776.327828] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] self.force_reraise() [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] raise self.value [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] updated_port = self._update_port( [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] _ensure_no_port_binding_failure(port) [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] raise exception.PortBindingFailed(port_id=port['id']) [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] nova.exception.PortBindingFailed: Binding failed for port a3482ef0-da11-44f3-84af-7490a7523ddd, please check neutron logs for more information. [ 776.328251] env[63355]: ERROR nova.compute.manager [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] [ 776.328600] env[63355]: DEBUG nova.compute.utils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Binding failed for port a3482ef0-da11-44f3-84af-7490a7523ddd, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 776.328981] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.893s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.331927] env[63355]: INFO nova.compute.claims [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.335138] env[63355]: DEBUG nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Build of instance 9dd8b658-886a-4f11-ab9b-059c4265436c was re-scheduled: Binding failed for port a3482ef0-da11-44f3-84af-7490a7523ddd, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 776.335660] env[63355]: DEBUG nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 776.336093] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "refresh_cache-9dd8b658-886a-4f11-ab9b-059c4265436c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.336294] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquired lock "refresh_cache-9dd8b658-886a-4f11-ab9b-059c4265436c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.336465] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 776.859071] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.953140] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.054625] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "b4b09b1d-680e-47b8-aa8a-9b3d9167824d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.054817] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "b4b09b1d-680e-47b8-aa8a-9b3d9167824d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.456541] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Releasing lock "refresh_cache-9dd8b658-886a-4f11-ab9b-059c4265436c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.456807] env[63355]: DEBUG nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 777.456959] env[63355]: DEBUG nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.457166] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 777.480057] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.742937] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a8200e-c472-4561-ae9f-4e6aa5e20c81 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.750937] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4df640-3d22-4d4c-a8fe-fd0596170103 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.782575] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ac6f38-b762-470d-9bd3-b6ecdf6409fa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.789613] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608ccce8-04a3-46a1-8526-2ae2e9e058d5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.802619] env[63355]: DEBUG nova.compute.provider_tree [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.983101] env[63355]: DEBUG nova.network.neutron [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.308185] env[63355]: DEBUG nova.scheduler.client.report [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.489744] env[63355]: INFO nova.compute.manager [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 9dd8b658-886a-4f11-ab9b-059c4265436c] Took 1.03 seconds to deallocate network for instance. [ 778.813013] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.813534] env[63355]: DEBUG nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 778.816932] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.047s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.323022] env[63355]: DEBUG nova.compute.utils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.323022] env[63355]: DEBUG nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 779.323022] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 779.399813] env[63355]: DEBUG nova.policy [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e420329c3714636bac314f5abf59f9b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b86199795d048c2b32852520d504f17', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.517092] env[63355]: INFO nova.scheduler.client.report [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Deleted allocations for instance 9dd8b658-886a-4f11-ab9b-059c4265436c [ 779.793917] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f12967-9838-493c-b605-b07b894e2605 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.802784] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff60ebc-be17-45a2-ae15-5069b6d4ec5f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.838267] env[63355]: DEBUG nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 779.842193] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802d9ad2-81c5-4baa-a08c-d920707f33a3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.850135] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8773e04-a85b-44c3-b513-e1673b673465 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.864276] env[63355]: DEBUG nova.compute.provider_tree [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.009909] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Successfully created port: 15746d33-dfc2-452b-a522-8cf17fbec8c0 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.024703] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aec1fa68-2cff-4e5b-afe4-b65fe74de04b tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "9dd8b658-886a-4f11-ab9b-059c4265436c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.399s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.370822] env[63355]: DEBUG nova.scheduler.client.report [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.527615] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 780.851582] env[63355]: DEBUG nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 780.876108] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.059s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.876854] env[63355]: ERROR nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a5cfc749-0746-45b0-b6b9-c9f92637b2d1, please check neutron logs for more information. [ 780.876854] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Traceback (most recent call last): [ 780.876854] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 780.876854] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] self.driver.spawn(context, instance, image_meta, [ 780.876854] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 780.876854] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] self._vmops.spawn(context, instance, image_meta, injected_files, [ 780.876854] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 780.876854] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] vm_ref = self.build_virtual_machine(instance, [ 780.876854] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 780.876854] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] vif_infos = vmwarevif.get_vif_info(self._session, [ 780.876854] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] for vif in network_info: [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] return self._sync_wrapper(fn, *args, **kwargs) [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] self.wait() [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] self[:] = self._gt.wait() [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] return self._exit_event.wait() [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] result = hub.switch() [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 780.880081] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] return self.greenlet.switch() [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] result = function(*args, **kwargs) [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] return func(*args, **kwargs) [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] raise e [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] nwinfo = self.network_api.allocate_for_instance( [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] created_port_ids = self._update_ports_for_instance( [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] with excutils.save_and_reraise_exception(): [ 780.880865] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] self.force_reraise() [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] raise self.value [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] updated_port = self._update_port( [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] _ensure_no_port_binding_failure(port) [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] raise exception.PortBindingFailed(port_id=port['id']) [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] nova.exception.PortBindingFailed: Binding failed for port a5cfc749-0746-45b0-b6b9-c9f92637b2d1, please check neutron logs for more information. [ 780.881328] env[63355]: ERROR nova.compute.manager [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] [ 780.881627] env[63355]: DEBUG nova.compute.utils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Binding failed for port a5cfc749-0746-45b0-b6b9-c9f92637b2d1, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 780.881627] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 780.881627] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 780.882613] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.882613] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 780.882613] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.882613] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 780.882613] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 780.882613] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 780.882840] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 780.882840] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 780.882840] env[63355]: DEBUG nova.virt.hardware [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.882930] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.608s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.886528] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587c62bf-9f12-4ed5-8a69-c2272f55b9ba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.894247] env[63355]: DEBUG nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Build of instance d7589681-fe6c-4dc6-b533-ade8ae634d58 was re-scheduled: Binding failed for port a5cfc749-0746-45b0-b6b9-c9f92637b2d1, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 780.894247] env[63355]: DEBUG nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 780.894247] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Acquiring lock "refresh_cache-d7589681-fe6c-4dc6-b533-ade8ae634d58" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.894247] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Acquired lock "refresh_cache-d7589681-fe6c-4dc6-b533-ade8ae634d58" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.894463] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.915066] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1c8214-a232-4b68-a216-6705673d14d4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.051540] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.102116] env[63355]: DEBUG nova.compute.manager [req-7a194e65-c831-4eed-b22e-470fd739e3ce req-426055f7-823e-4426-9a8a-99b808869c64 service nova] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Received event network-changed-15746d33-dfc2-452b-a522-8cf17fbec8c0 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.102116] env[63355]: DEBUG nova.compute.manager [req-7a194e65-c831-4eed-b22e-470fd739e3ce req-426055f7-823e-4426-9a8a-99b808869c64 service nova] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Refreshing instance network info cache due to event network-changed-15746d33-dfc2-452b-a522-8cf17fbec8c0. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 781.102116] env[63355]: DEBUG oslo_concurrency.lockutils [req-7a194e65-c831-4eed-b22e-470fd739e3ce req-426055f7-823e-4426-9a8a-99b808869c64 service nova] Acquiring lock "refresh_cache-ec4726b2-38f9-4499-a5b0-a70db0d697f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.102116] env[63355]: DEBUG oslo_concurrency.lockutils [req-7a194e65-c831-4eed-b22e-470fd739e3ce req-426055f7-823e-4426-9a8a-99b808869c64 service nova] Acquired lock "refresh_cache-ec4726b2-38f9-4499-a5b0-a70db0d697f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.102116] env[63355]: DEBUG nova.network.neutron [req-7a194e65-c831-4eed-b22e-470fd739e3ce req-426055f7-823e-4426-9a8a-99b808869c64 service nova] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Refreshing network info cache for port 15746d33-dfc2-452b-a522-8cf17fbec8c0 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 781.177258] env[63355]: ERROR nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 15746d33-dfc2-452b-a522-8cf17fbec8c0, please check neutron logs for more information. [ 781.177258] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 781.177258] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.177258] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 781.177258] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.177258] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 781.177258] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.177258] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 781.177258] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.177258] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 781.177258] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.177258] env[63355]: ERROR nova.compute.manager raise self.value [ 781.177258] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.177258] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 781.177258] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.177258] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 781.177793] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.177793] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 781.177793] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 15746d33-dfc2-452b-a522-8cf17fbec8c0, please check neutron logs for more information. [ 781.177793] env[63355]: ERROR nova.compute.manager [ 781.177793] env[63355]: Traceback (most recent call last): [ 781.177793] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 781.177793] env[63355]: listener.cb(fileno) [ 781.177793] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.177793] env[63355]: result = function(*args, **kwargs) [ 781.177793] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 781.177793] env[63355]: return func(*args, **kwargs) [ 781.177793] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.177793] env[63355]: raise e [ 781.177793] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.177793] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 781.177793] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.177793] env[63355]: created_port_ids = self._update_ports_for_instance( [ 781.177793] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.177793] env[63355]: with excutils.save_and_reraise_exception(): [ 781.177793] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.177793] env[63355]: self.force_reraise() [ 781.177793] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.177793] env[63355]: raise self.value [ 781.177793] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.177793] env[63355]: updated_port = self._update_port( [ 781.177793] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.177793] env[63355]: _ensure_no_port_binding_failure(port) [ 781.177793] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.177793] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 781.178671] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 15746d33-dfc2-452b-a522-8cf17fbec8c0, please check neutron logs for more information. [ 781.178671] env[63355]: Removing descriptor: 17 [ 781.178671] env[63355]: ERROR nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 15746d33-dfc2-452b-a522-8cf17fbec8c0, please check neutron logs for more information. [ 781.178671] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Traceback (most recent call last): [ 781.178671] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 781.178671] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] yield resources [ 781.178671] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.178671] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] self.driver.spawn(context, instance, image_meta, [ 781.178671] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 781.178671] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.178671] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.178671] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] vm_ref = self.build_virtual_machine(instance, [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] for vif in network_info: [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] return self._sync_wrapper(fn, *args, **kwargs) [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] self.wait() [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] self[:] = self._gt.wait() [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] return self._exit_event.wait() [ 781.179151] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] result = hub.switch() [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] return self.greenlet.switch() [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] result = function(*args, **kwargs) [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] return func(*args, **kwargs) [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] raise e [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] nwinfo = self.network_api.allocate_for_instance( [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.179564] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] created_port_ids = self._update_ports_for_instance( [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] with excutils.save_and_reraise_exception(): [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] self.force_reraise() [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] raise self.value [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] updated_port = self._update_port( [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] _ensure_no_port_binding_failure(port) [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.179969] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] raise exception.PortBindingFailed(port_id=port['id']) [ 781.180322] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] nova.exception.PortBindingFailed: Binding failed for port 15746d33-dfc2-452b-a522-8cf17fbec8c0, please check neutron logs for more information. [ 781.180322] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] [ 781.180322] env[63355]: INFO nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Terminating instance [ 781.180837] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Acquiring lock "refresh_cache-ec4726b2-38f9-4499-a5b0-a70db0d697f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.431493] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.521229] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.626843] env[63355]: DEBUG nova.network.neutron [req-7a194e65-c831-4eed-b22e-470fd739e3ce req-426055f7-823e-4426-9a8a-99b808869c64 service nova] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.704799] env[63355]: DEBUG nova.network.neutron [req-7a194e65-c831-4eed-b22e-470fd739e3ce req-426055f7-823e-4426-9a8a-99b808869c64 service nova] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.796149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquiring lock "2539a79e-01c0-4e0c-aa66-8784441c6fda" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.796149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lock "2539a79e-01c0-4e0c-aa66-8784441c6fda" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.799470] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b46e696-75cc-4dd4-b67a-99adda405f13 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.807900] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490f7ed2-e37c-42b5-a29b-93d781d96a53 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.838673] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaef6f81-d7c6-47d7-a126-24ee18e575d0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.845955] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba39623-1517-4fb0-8d0c-f96815144706 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.858882] env[63355]: DEBUG nova.compute.provider_tree [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.023723] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Releasing lock "refresh_cache-d7589681-fe6c-4dc6-b533-ade8ae634d58" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.023940] env[63355]: DEBUG nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.024149] env[63355]: DEBUG nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.024320] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.040824] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.207219] env[63355]: DEBUG oslo_concurrency.lockutils [req-7a194e65-c831-4eed-b22e-470fd739e3ce req-426055f7-823e-4426-9a8a-99b808869c64 service nova] Releasing lock "refresh_cache-ec4726b2-38f9-4499-a5b0-a70db0d697f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.207620] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Acquired lock "refresh_cache-ec4726b2-38f9-4499-a5b0-a70db0d697f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.207821] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.362709] env[63355]: DEBUG nova.scheduler.client.report [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.543392] env[63355]: DEBUG nova.network.neutron [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.725400] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.803114] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.867256] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.867867] env[63355]: ERROR nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0f896ea1-ac3c-418a-9a64-5efa608ee2bd, please check neutron logs for more information. [ 782.867867] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Traceback (most recent call last): [ 782.867867] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 782.867867] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] self.driver.spawn(context, instance, image_meta, [ 782.867867] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 782.867867] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] self._vmops.spawn(context, instance, image_meta, injected_files, [ 782.867867] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 782.867867] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] vm_ref = self.build_virtual_machine(instance, [ 782.867867] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 782.867867] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] vif_infos = vmwarevif.get_vif_info(self._session, [ 782.867867] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] for vif in network_info: [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] return self._sync_wrapper(fn, *args, **kwargs) [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] self.wait() [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] self[:] = self._gt.wait() [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] return self._exit_event.wait() [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] result = hub.switch() [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 782.868244] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] return self.greenlet.switch() [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] result = function(*args, **kwargs) [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] return func(*args, **kwargs) [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] raise e [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] nwinfo = self.network_api.allocate_for_instance( [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] created_port_ids = self._update_ports_for_instance( [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] with excutils.save_and_reraise_exception(): [ 782.868655] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] self.force_reraise() [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] raise self.value [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] updated_port = self._update_port( [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] _ensure_no_port_binding_failure(port) [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] raise exception.PortBindingFailed(port_id=port['id']) [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] nova.exception.PortBindingFailed: Binding failed for port 0f896ea1-ac3c-418a-9a64-5efa608ee2bd, please check neutron logs for more information. [ 782.869080] env[63355]: ERROR nova.compute.manager [instance: 17a904fe-ab4b-46ed-98c8-175987710643] [ 782.869424] env[63355]: DEBUG nova.compute.utils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Binding failed for port 0f896ea1-ac3c-418a-9a64-5efa608ee2bd, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 782.869701] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.494s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.871155] env[63355]: INFO nova.compute.claims [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.873617] env[63355]: DEBUG nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Build of instance 17a904fe-ab4b-46ed-98c8-175987710643 was re-scheduled: Binding failed for port 0f896ea1-ac3c-418a-9a64-5efa608ee2bd, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 782.875179] env[63355]: DEBUG nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 782.875179] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "refresh_cache-17a904fe-ab4b-46ed-98c8-175987710643" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.875179] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquired lock "refresh_cache-17a904fe-ab4b-46ed-98c8-175987710643" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.875179] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.045858] env[63355]: INFO nova.compute.manager [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] [instance: d7589681-fe6c-4dc6-b533-ade8ae634d58] Took 1.02 seconds to deallocate network for instance. [ 783.126386] env[63355]: DEBUG nova.compute.manager [req-8ae26123-3083-4fae-8834-a9ea7cec7640 req-e664d0a4-cf86-41b6-a850-05b1ccd56133 service nova] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Received event network-vif-deleted-15746d33-dfc2-452b-a522-8cf17fbec8c0 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 783.305704] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Releasing lock "refresh_cache-ec4726b2-38f9-4499-a5b0-a70db0d697f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.306159] env[63355]: DEBUG nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 783.306392] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 783.306739] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-31b8c3e7-0c40-4f6d-b0da-357aac1b2c36 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.316194] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54721a72-b424-4b77-800e-4a45d72fd70f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.338058] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ec4726b2-38f9-4499-a5b0-a70db0d697f6 could not be found. [ 783.338287] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 783.338469] env[63355]: INFO nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 783.338689] env[63355]: DEBUG oslo.service.loopingcall [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.338904] env[63355]: DEBUG nova.compute.manager [-] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.338996] env[63355]: DEBUG nova.network.neutron [-] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 783.354539] env[63355]: DEBUG nova.network.neutron [-] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.393046] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.460531] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.857131] env[63355]: DEBUG nova.network.neutron [-] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.963575] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Releasing lock "refresh_cache-17a904fe-ab4b-46ed-98c8-175987710643" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.963868] env[63355]: DEBUG nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 783.964092] env[63355]: DEBUG nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.964270] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 783.981027] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.071867] env[63355]: INFO nova.scheduler.client.report [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Deleted allocations for instance d7589681-fe6c-4dc6-b533-ade8ae634d58 [ 784.234056] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7978fdda-b308-47a9-9b0d-a285983401eb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.242023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e35de9-b185-4858-a904-e245d08043a6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.271751] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a631e9ff-f742-4664-b8fe-cec5dd5c1eb3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.278027] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c501dc0e-1d94-4389-b94b-5bdd5a8f7104 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.291749] env[63355]: DEBUG nova.compute.provider_tree [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.360297] env[63355]: INFO nova.compute.manager [-] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Took 1.02 seconds to deallocate network for instance. [ 784.363691] env[63355]: DEBUG nova.compute.claims [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 784.363880] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.483409] env[63355]: DEBUG nova.network.neutron [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.582022] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b3cde40c-86b8-48cd-a560-0c2db7273b0f tempest-ServersV294TestFqdnHostnames-452651400 tempest-ServersV294TestFqdnHostnames-452651400-project-member] Lock "d7589681-fe6c-4dc6-b533-ade8ae634d58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.925s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.795024] env[63355]: DEBUG nova.scheduler.client.report [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.986397] env[63355]: INFO nova.compute.manager [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: 17a904fe-ab4b-46ed-98c8-175987710643] Took 1.02 seconds to deallocate network for instance. [ 785.085202] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.302025] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.302025] env[63355]: DEBUG nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.302884] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.940s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.304445] env[63355]: INFO nova.compute.claims [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.621923] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.808981] env[63355]: DEBUG nova.compute.utils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.813198] env[63355]: DEBUG nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.813440] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 785.881910] env[63355]: DEBUG nova.policy [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3813b6016e9c4b2b8fdf7a2db1a8a74b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81adc0ecb7dd4c2fa22110ca4f54d396', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 786.023453] env[63355]: INFO nova.scheduler.client.report [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Deleted allocations for instance 17a904fe-ab4b-46ed-98c8-175987710643 [ 786.313811] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Successfully created port: 34d3c5d5-253d-4236-9f68-078074585482 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.315638] env[63355]: DEBUG nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.536467] env[63355]: DEBUG oslo_concurrency.lockutils [None req-aa992d71-18b4-41c0-90e8-75536801ff64 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "17a904fe-ab4b-46ed-98c8-175987710643" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.955s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.770740] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91cffa29-c931-4f11-bcbf-3ca90d2b3196 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.778519] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90cf9708-3fbc-4ee5-a859-953199dbf3db {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.810330] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e48477-16d5-4514-981c-6a7691761303 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.818854] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd015189-66a9-4c8a-8219-4fda53fb39c7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.834128] env[63355]: DEBUG nova.compute.provider_tree [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.039974] env[63355]: DEBUG nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 787.221766] env[63355]: DEBUG nova.compute.manager [req-22414113-45aa-4ded-8ce8-039f2eaf3bef req-c50f86ce-5549-4e18-9e6b-72ed46f261b3 service nova] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Received event network-changed-34d3c5d5-253d-4236-9f68-078074585482 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.221962] env[63355]: DEBUG nova.compute.manager [req-22414113-45aa-4ded-8ce8-039f2eaf3bef req-c50f86ce-5549-4e18-9e6b-72ed46f261b3 service nova] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Refreshing instance network info cache due to event network-changed-34d3c5d5-253d-4236-9f68-078074585482. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 787.222190] env[63355]: DEBUG oslo_concurrency.lockutils [req-22414113-45aa-4ded-8ce8-039f2eaf3bef req-c50f86ce-5549-4e18-9e6b-72ed46f261b3 service nova] Acquiring lock "refresh_cache-3c04c546-a9fa-4cbd-98d4-3ea92918a918" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.222329] env[63355]: DEBUG oslo_concurrency.lockutils [req-22414113-45aa-4ded-8ce8-039f2eaf3bef req-c50f86ce-5549-4e18-9e6b-72ed46f261b3 service nova] Acquired lock "refresh_cache-3c04c546-a9fa-4cbd-98d4-3ea92918a918" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.222518] env[63355]: DEBUG nova.network.neutron [req-22414113-45aa-4ded-8ce8-039f2eaf3bef req-c50f86ce-5549-4e18-9e6b-72ed46f261b3 service nova] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Refreshing network info cache for port 34d3c5d5-253d-4236-9f68-078074585482 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 787.337590] env[63355]: DEBUG nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.340715] env[63355]: DEBUG nova.scheduler.client.report [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.378609] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.378609] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.378609] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.378787] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.378787] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.378860] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.379266] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.379484] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.379714] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.379907] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.380117] env[63355]: DEBUG nova.virt.hardware [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.381344] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62aeffc0-f62c-410c-bfb1-affe2641ca2d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.389825] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ac795e-7b80-42da-ad88-cc61cf64b33e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.481203] env[63355]: ERROR nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 34d3c5d5-253d-4236-9f68-078074585482, please check neutron logs for more information. [ 787.481203] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 787.481203] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.481203] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 787.481203] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.481203] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 787.481203] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.481203] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 787.481203] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.481203] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 787.481203] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.481203] env[63355]: ERROR nova.compute.manager raise self.value [ 787.481203] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.481203] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 787.481203] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.481203] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 787.481761] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.481761] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 787.481761] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 34d3c5d5-253d-4236-9f68-078074585482, please check neutron logs for more information. [ 787.481761] env[63355]: ERROR nova.compute.manager [ 787.481761] env[63355]: Traceback (most recent call last): [ 787.481761] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 787.481761] env[63355]: listener.cb(fileno) [ 787.481761] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.481761] env[63355]: result = function(*args, **kwargs) [ 787.481761] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 787.481761] env[63355]: return func(*args, **kwargs) [ 787.481761] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.481761] env[63355]: raise e [ 787.481761] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.481761] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 787.481761] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.481761] env[63355]: created_port_ids = self._update_ports_for_instance( [ 787.481761] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.481761] env[63355]: with excutils.save_and_reraise_exception(): [ 787.481761] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.481761] env[63355]: self.force_reraise() [ 787.481761] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.481761] env[63355]: raise self.value [ 787.481761] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.481761] env[63355]: updated_port = self._update_port( [ 787.481761] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.481761] env[63355]: _ensure_no_port_binding_failure(port) [ 787.481761] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.481761] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 787.482762] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 34d3c5d5-253d-4236-9f68-078074585482, please check neutron logs for more information. [ 787.482762] env[63355]: Removing descriptor: 17 [ 787.482762] env[63355]: ERROR nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 34d3c5d5-253d-4236-9f68-078074585482, please check neutron logs for more information. [ 787.482762] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Traceback (most recent call last): [ 787.482762] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 787.482762] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] yield resources [ 787.482762] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 787.482762] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] self.driver.spawn(context, instance, image_meta, [ 787.482762] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 787.482762] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.482762] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.482762] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] vm_ref = self.build_virtual_machine(instance, [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] for vif in network_info: [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] return self._sync_wrapper(fn, *args, **kwargs) [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] self.wait() [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] self[:] = self._gt.wait() [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] return self._exit_event.wait() [ 787.483161] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] result = hub.switch() [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] return self.greenlet.switch() [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] result = function(*args, **kwargs) [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] return func(*args, **kwargs) [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] raise e [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] nwinfo = self.network_api.allocate_for_instance( [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.483577] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] created_port_ids = self._update_ports_for_instance( [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] with excutils.save_and_reraise_exception(): [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] self.force_reraise() [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] raise self.value [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] updated_port = self._update_port( [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] _ensure_no_port_binding_failure(port) [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.484013] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] raise exception.PortBindingFailed(port_id=port['id']) [ 787.484399] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] nova.exception.PortBindingFailed: Binding failed for port 34d3c5d5-253d-4236-9f68-078074585482, please check neutron logs for more information. [ 787.484399] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] [ 787.484399] env[63355]: INFO nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Terminating instance [ 787.485822] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Acquiring lock "refresh_cache-3c04c546-a9fa-4cbd-98d4-3ea92918a918" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.562968] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.746231] env[63355]: DEBUG nova.network.neutron [req-22414113-45aa-4ded-8ce8-039f2eaf3bef req-c50f86ce-5549-4e18-9e6b-72ed46f261b3 service nova] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.846089] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.846647] env[63355]: DEBUG nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.851618] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.500s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.853188] env[63355]: INFO nova.compute.claims [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.866372] env[63355]: DEBUG nova.network.neutron [req-22414113-45aa-4ded-8ce8-039f2eaf3bef req-c50f86ce-5549-4e18-9e6b-72ed46f261b3 service nova] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.362030] env[63355]: DEBUG nova.compute.utils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.369194] env[63355]: DEBUG nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 788.369452] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 788.371851] env[63355]: DEBUG oslo_concurrency.lockutils [req-22414113-45aa-4ded-8ce8-039f2eaf3bef req-c50f86ce-5549-4e18-9e6b-72ed46f261b3 service nova] Releasing lock "refresh_cache-3c04c546-a9fa-4cbd-98d4-3ea92918a918" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.372248] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Acquired lock "refresh_cache-3c04c546-a9fa-4cbd-98d4-3ea92918a918" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.372423] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.437876] env[63355]: DEBUG nova.policy [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '00d6e1e0ce804255b530760d8f959261', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8354011bc12a44779eb4030247748227', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 788.866270] env[63355]: DEBUG nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.869731] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Successfully created port: fcd0ef62-5937-42f8-ae04-94e0e375456e {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.902219] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.045817] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.332879] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17337938-1914-4ee3-b019-c02b0a0adbee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.342823] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8886aea-3e1a-44a1-ae56-834a48ee2d84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.378248] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6efd22-dd9b-4483-96d1-1febed8e8698 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.387480] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ee8a87-f90f-4389-88fb-6932ce371d12 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.408053] env[63355]: DEBUG nova.compute.provider_tree [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.491453] env[63355]: DEBUG nova.compute.manager [req-16b20ab0-0c71-4156-8039-ec5142df4264 req-26f212ba-1f16-480d-9be6-864fd4b16dcb service nova] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Received event network-vif-deleted-34d3c5d5-253d-4236-9f68-078074585482 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.558808] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Releasing lock "refresh_cache-3c04c546-a9fa-4cbd-98d4-3ea92918a918" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.558808] env[63355]: DEBUG nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 789.558808] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.558808] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-718fc977-b48c-4f04-ab84-8d1ab3daac19 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.568634] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392d1e0d-5ff0-4868-9d0b-f55850501f48 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.594030] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3c04c546-a9fa-4cbd-98d4-3ea92918a918 could not be found. [ 789.594152] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.594274] env[63355]: INFO nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Took 0.04 seconds to destroy the instance on the hypervisor. [ 789.594522] env[63355]: DEBUG oslo.service.loopingcall [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.594975] env[63355]: DEBUG nova.compute.manager [-] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.595119] env[63355]: DEBUG nova.network.neutron [-] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 789.615258] env[63355]: DEBUG nova.network.neutron [-] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.884851] env[63355]: DEBUG nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 789.912498] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.912747] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.912896] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.913088] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.913243] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.913392] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.913595] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.913754] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.913916] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.914240] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.914449] env[63355]: DEBUG nova.virt.hardware [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.917214] env[63355]: DEBUG nova.scheduler.client.report [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.922746] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540c2942-d843-4c9c-b511-f41f891d5526 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.927671] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082ab830-ad82-465b-9985-262f218603b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.121264] env[63355]: DEBUG nova.network.neutron [-] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.266929] env[63355]: ERROR nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fcd0ef62-5937-42f8-ae04-94e0e375456e, please check neutron logs for more information. [ 790.266929] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 790.266929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.266929] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 790.266929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.266929] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 790.266929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.266929] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 790.266929] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.266929] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 790.266929] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.266929] env[63355]: ERROR nova.compute.manager raise self.value [ 790.266929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.266929] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 790.266929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.266929] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 790.267459] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.267459] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 790.267459] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fcd0ef62-5937-42f8-ae04-94e0e375456e, please check neutron logs for more information. [ 790.267459] env[63355]: ERROR nova.compute.manager [ 790.267459] env[63355]: Traceback (most recent call last): [ 790.267459] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 790.267459] env[63355]: listener.cb(fileno) [ 790.267459] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.267459] env[63355]: result = function(*args, **kwargs) [ 790.267459] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 790.267459] env[63355]: return func(*args, **kwargs) [ 790.267459] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.267459] env[63355]: raise e [ 790.267459] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.267459] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 790.267459] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.267459] env[63355]: created_port_ids = self._update_ports_for_instance( [ 790.267459] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.267459] env[63355]: with excutils.save_and_reraise_exception(): [ 790.267459] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.267459] env[63355]: self.force_reraise() [ 790.267459] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.267459] env[63355]: raise self.value [ 790.267459] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.267459] env[63355]: updated_port = self._update_port( [ 790.267459] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.267459] env[63355]: _ensure_no_port_binding_failure(port) [ 790.267459] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.267459] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 790.268484] env[63355]: nova.exception.PortBindingFailed: Binding failed for port fcd0ef62-5937-42f8-ae04-94e0e375456e, please check neutron logs for more information. [ 790.268484] env[63355]: Removing descriptor: 17 [ 790.268484] env[63355]: ERROR nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fcd0ef62-5937-42f8-ae04-94e0e375456e, please check neutron logs for more information. [ 790.268484] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Traceback (most recent call last): [ 790.268484] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 790.268484] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] yield resources [ 790.268484] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 790.268484] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] self.driver.spawn(context, instance, image_meta, [ 790.268484] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 790.268484] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.268484] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.268484] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] vm_ref = self.build_virtual_machine(instance, [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] vif_infos = vmwarevif.get_vif_info(self._session, [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] for vif in network_info: [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] return self._sync_wrapper(fn, *args, **kwargs) [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] self.wait() [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] self[:] = self._gt.wait() [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] return self._exit_event.wait() [ 790.268902] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] result = hub.switch() [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] return self.greenlet.switch() [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] result = function(*args, **kwargs) [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] return func(*args, **kwargs) [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] raise e [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] nwinfo = self.network_api.allocate_for_instance( [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.269343] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] created_port_ids = self._update_ports_for_instance( [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] with excutils.save_and_reraise_exception(): [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] self.force_reraise() [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] raise self.value [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] updated_port = self._update_port( [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] _ensure_no_port_binding_failure(port) [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.269781] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] raise exception.PortBindingFailed(port_id=port['id']) [ 790.270195] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] nova.exception.PortBindingFailed: Binding failed for port fcd0ef62-5937-42f8-ae04-94e0e375456e, please check neutron logs for more information. [ 790.270195] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] [ 790.270195] env[63355]: INFO nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Terminating instance [ 790.270301] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Acquiring lock "refresh_cache-05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.270431] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Acquired lock "refresh_cache-05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.270570] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.424984] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.576s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.425455] env[63355]: DEBUG nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 790.429226] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.328s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.625035] env[63355]: INFO nova.compute.manager [-] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Took 1.03 seconds to deallocate network for instance. [ 790.627665] env[63355]: DEBUG nova.compute.claims [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 790.627665] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.789669] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.864283] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.931623] env[63355]: DEBUG nova.compute.utils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.933131] env[63355]: DEBUG nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.933342] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 790.991969] env[63355]: DEBUG nova.policy [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2681de85a8d944529976b42dd02367b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '793fe76e562d4e3a81e3ed2f2d6ee261', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 791.115856] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquiring lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.115856] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.292758] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Successfully created port: 260b6560-883f-441f-ada4-45fef2d173d6 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.351131] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f600f7-7986-4369-a3f0-9846b77000e6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.360068] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9148c8f9-0005-48aa-86c7-0a4b553b2f31 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.367027] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Releasing lock "refresh_cache-05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.367776] env[63355]: DEBUG nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 791.367776] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 791.368074] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-362f798b-ddad-4dc0-a7ff-288bd8a49e60 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.396668] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc76d7f3-5be7-4433-86c9-e0802992b665 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.402849] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6439ab-eba0-4edb-b0f5-7aecfc662e8b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.415786] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1c7acb-b439-4121-add9-b26b70d44559 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.430107] env[63355]: DEBUG nova.compute.provider_tree [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.439263] env[63355]: DEBUG nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 791.447155] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e could not be found. [ 791.447155] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.447155] env[63355]: INFO nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Took 0.08 seconds to destroy the instance on the hypervisor. [ 791.447155] env[63355]: DEBUG oslo.service.loopingcall [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.447155] env[63355]: DEBUG nova.compute.manager [-] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.447155] env[63355]: DEBUG nova.network.neutron [-] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.469580] env[63355]: DEBUG nova.network.neutron [-] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.531514] env[63355]: DEBUG nova.compute.manager [req-055a49f0-6ef0-4a98-9a85-d03a5e8ae73f req-86e26a9d-7192-424f-80ec-6ba4fed75f22 service nova] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Received event network-changed-fcd0ef62-5937-42f8-ae04-94e0e375456e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.531777] env[63355]: DEBUG nova.compute.manager [req-055a49f0-6ef0-4a98-9a85-d03a5e8ae73f req-86e26a9d-7192-424f-80ec-6ba4fed75f22 service nova] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Refreshing instance network info cache due to event network-changed-fcd0ef62-5937-42f8-ae04-94e0e375456e. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 791.532020] env[63355]: DEBUG oslo_concurrency.lockutils [req-055a49f0-6ef0-4a98-9a85-d03a5e8ae73f req-86e26a9d-7192-424f-80ec-6ba4fed75f22 service nova] Acquiring lock "refresh_cache-05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.532165] env[63355]: DEBUG oslo_concurrency.lockutils [req-055a49f0-6ef0-4a98-9a85-d03a5e8ae73f req-86e26a9d-7192-424f-80ec-6ba4fed75f22 service nova] Acquired lock "refresh_cache-05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.532320] env[63355]: DEBUG nova.network.neutron [req-055a49f0-6ef0-4a98-9a85-d03a5e8ae73f req-86e26a9d-7192-424f-80ec-6ba4fed75f22 service nova] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Refreshing network info cache for port fcd0ef62-5937-42f8-ae04-94e0e375456e {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 791.949117] env[63355]: DEBUG nova.scheduler.client.report [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.957093] env[63355]: INFO nova.virt.block_device [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Booting with volume 56f0283a-f2fa-4ff2-a7fc-4af53c2f3f7d at /dev/sda [ 791.978887] env[63355]: DEBUG nova.network.neutron [-] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.023768] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4274674-fbe9-444f-b65b-446ea998a402 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.034376] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45f5c84-eb94-4459-8d2f-a06687b97e87 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.060084] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81f242f9-5e77-4306-a80e-64fd8366f8f9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.068063] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033e1ed6-9f05-4683-bcd3-6b7d4b120431 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.079877] env[63355]: DEBUG nova.network.neutron [req-055a49f0-6ef0-4a98-9a85-d03a5e8ae73f req-86e26a9d-7192-424f-80ec-6ba4fed75f22 service nova] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.092730] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e85da31-845a-4174-bee2-5da6f4a2568e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.099539] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55d5244-81df-48ae-9cc0-b78c2d4be530 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.119773] env[63355]: DEBUG nova.virt.block_device [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Updating existing volume attachment record: 21696c34-667d-4fee-9979-463d01ea3268 {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 792.213579] env[63355]: DEBUG nova.network.neutron [req-055a49f0-6ef0-4a98-9a85-d03a5e8ae73f req-86e26a9d-7192-424f-80ec-6ba4fed75f22 service nova] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.217186] env[63355]: ERROR nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 260b6560-883f-441f-ada4-45fef2d173d6, please check neutron logs for more information. [ 792.217186] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 792.217186] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.217186] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 792.217186] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.217186] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 792.217186] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.217186] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 792.217186] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.217186] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 792.217186] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.217186] env[63355]: ERROR nova.compute.manager raise self.value [ 792.217186] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.217186] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 792.217186] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.217186] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 792.217683] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.217683] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 792.217683] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 260b6560-883f-441f-ada4-45fef2d173d6, please check neutron logs for more information. [ 792.217683] env[63355]: ERROR nova.compute.manager [ 792.217683] env[63355]: Traceback (most recent call last): [ 792.217683] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 792.217683] env[63355]: listener.cb(fileno) [ 792.217683] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.217683] env[63355]: result = function(*args, **kwargs) [ 792.217683] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 792.217683] env[63355]: return func(*args, **kwargs) [ 792.217683] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.217683] env[63355]: raise e [ 792.217683] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.217683] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 792.217683] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.217683] env[63355]: created_port_ids = self._update_ports_for_instance( [ 792.217683] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.217683] env[63355]: with excutils.save_and_reraise_exception(): [ 792.217683] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.217683] env[63355]: self.force_reraise() [ 792.217683] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.217683] env[63355]: raise self.value [ 792.217683] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.217683] env[63355]: updated_port = self._update_port( [ 792.217683] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.217683] env[63355]: _ensure_no_port_binding_failure(port) [ 792.217683] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.217683] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 792.218636] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 260b6560-883f-441f-ada4-45fef2d173d6, please check neutron logs for more information. [ 792.218636] env[63355]: Removing descriptor: 17 [ 792.461715] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.033s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.462278] env[63355]: ERROR nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port aa866120-3950-4692-9cd4-afc27bb3a783, please check neutron logs for more information. [ 792.462278] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Traceback (most recent call last): [ 792.462278] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 792.462278] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] self.driver.spawn(context, instance, image_meta, [ 792.462278] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 792.462278] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 792.462278] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 792.462278] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] vm_ref = self.build_virtual_machine(instance, [ 792.462278] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 792.462278] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 792.462278] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] for vif in network_info: [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] return self._sync_wrapper(fn, *args, **kwargs) [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] self.wait() [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] self[:] = self._gt.wait() [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] return self._exit_event.wait() [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] current.throw(*self._exc) [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.462691] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] result = function(*args, **kwargs) [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] return func(*args, **kwargs) [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] raise e [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] nwinfo = self.network_api.allocate_for_instance( [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] created_port_ids = self._update_ports_for_instance( [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] with excutils.save_and_reraise_exception(): [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] self.force_reraise() [ 792.463119] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.463620] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] raise self.value [ 792.463620] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.463620] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] updated_port = self._update_port( [ 792.463620] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.463620] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] _ensure_no_port_binding_failure(port) [ 792.463620] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.463620] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] raise exception.PortBindingFailed(port_id=port['id']) [ 792.463620] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] nova.exception.PortBindingFailed: Binding failed for port aa866120-3950-4692-9cd4-afc27bb3a783, please check neutron logs for more information. [ 792.463620] env[63355]: ERROR nova.compute.manager [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] [ 792.463620] env[63355]: DEBUG nova.compute.utils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Binding failed for port aa866120-3950-4692-9cd4-afc27bb3a783, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 792.464249] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.079s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.467682] env[63355]: DEBUG nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Build of instance d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6 was re-scheduled: Binding failed for port aa866120-3950-4692-9cd4-afc27bb3a783, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 792.468099] env[63355]: DEBUG nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 792.468322] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquiring lock "refresh_cache-d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.468477] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Acquired lock "refresh_cache-d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.468610] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.486746] env[63355]: INFO nova.compute.manager [-] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Took 1.04 seconds to deallocate network for instance. [ 792.488732] env[63355]: DEBUG nova.compute.claims [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 792.489127] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.715761] env[63355]: DEBUG oslo_concurrency.lockutils [req-055a49f0-6ef0-4a98-9a85-d03a5e8ae73f req-86e26a9d-7192-424f-80ec-6ba4fed75f22 service nova] Releasing lock "refresh_cache-05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.716124] env[63355]: DEBUG nova.compute.manager [req-055a49f0-6ef0-4a98-9a85-d03a5e8ae73f req-86e26a9d-7192-424f-80ec-6ba4fed75f22 service nova] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Received event network-vif-deleted-fcd0ef62-5937-42f8-ae04-94e0e375456e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.988923] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.075656] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.330029] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-797ef70f-9ffd-4b3b-b365-e8e9ede8741e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.337810] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34d62ed-a2ae-45dd-8434-98dc3ae1cbda {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.366423] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7c2e01-2136-488b-94e6-5e079fd6c14f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.373342] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac219b0-dd2e-4e77-b3d1-a4c6993e6249 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.387176] env[63355]: DEBUG nova.compute.provider_tree [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.557615] env[63355]: DEBUG nova.compute.manager [req-8f037691-6c4d-4407-bfb6-265abba14719 req-3f8021fb-5fc8-46d3-8cd1-25d0c9e1838d service nova] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Received event network-changed-260b6560-883f-441f-ada4-45fef2d173d6 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.557831] env[63355]: DEBUG nova.compute.manager [req-8f037691-6c4d-4407-bfb6-265abba14719 req-3f8021fb-5fc8-46d3-8cd1-25d0c9e1838d service nova] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Refreshing instance network info cache due to event network-changed-260b6560-883f-441f-ada4-45fef2d173d6. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.558103] env[63355]: DEBUG oslo_concurrency.lockutils [req-8f037691-6c4d-4407-bfb6-265abba14719 req-3f8021fb-5fc8-46d3-8cd1-25d0c9e1838d service nova] Acquiring lock "refresh_cache-282f1b55-8a74-467a-9ec8-0684eb780970" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.558190] env[63355]: DEBUG oslo_concurrency.lockutils [req-8f037691-6c4d-4407-bfb6-265abba14719 req-3f8021fb-5fc8-46d3-8cd1-25d0c9e1838d service nova] Acquired lock "refresh_cache-282f1b55-8a74-467a-9ec8-0684eb780970" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.558318] env[63355]: DEBUG nova.network.neutron [req-8f037691-6c4d-4407-bfb6-265abba14719 req-3f8021fb-5fc8-46d3-8cd1-25d0c9e1838d service nova] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Refreshing network info cache for port 260b6560-883f-441f-ada4-45fef2d173d6 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 793.579635] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Releasing lock "refresh_cache-d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.579851] env[63355]: DEBUG nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 793.580037] env[63355]: DEBUG nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.580203] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 793.725662] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.890050] env[63355]: DEBUG nova.scheduler.client.report [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.076624] env[63355]: DEBUG nova.network.neutron [req-8f037691-6c4d-4407-bfb6-265abba14719 req-3f8021fb-5fc8-46d3-8cd1-25d0c9e1838d service nova] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.150955] env[63355]: DEBUG nova.network.neutron [req-8f037691-6c4d-4407-bfb6-265abba14719 req-3f8021fb-5fc8-46d3-8cd1-25d0c9e1838d service nova] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.227785] env[63355]: DEBUG nova.network.neutron [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.240419] env[63355]: DEBUG nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 794.240940] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 794.241162] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 794.241315] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.241493] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 794.241636] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.241779] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 794.242052] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 794.242137] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 794.242302] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 794.242525] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 794.242683] env[63355]: DEBUG nova.virt.hardware [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 794.243932] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2754c43-1480-4071-9c03-c01adcf27ea8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.252666] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78055e0-3ebc-401a-8f72-db24f09588e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.266342] env[63355]: ERROR nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 260b6560-883f-441f-ada4-45fef2d173d6, please check neutron logs for more information. [ 794.266342] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Traceback (most recent call last): [ 794.266342] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 794.266342] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] yield resources [ 794.266342] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 794.266342] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] self.driver.spawn(context, instance, image_meta, [ 794.266342] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 794.266342] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] self._vmops.spawn(context, instance, image_meta, injected_files, [ 794.266342] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 794.266342] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] vm_ref = self.build_virtual_machine(instance, [ 794.266342] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] vif_infos = vmwarevif.get_vif_info(self._session, [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] for vif in network_info: [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] return self._sync_wrapper(fn, *args, **kwargs) [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] self.wait() [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] self[:] = self._gt.wait() [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] return self._exit_event.wait() [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 794.266737] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] current.throw(*self._exc) [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] result = function(*args, **kwargs) [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] return func(*args, **kwargs) [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] raise e [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] nwinfo = self.network_api.allocate_for_instance( [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] created_port_ids = self._update_ports_for_instance( [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] with excutils.save_and_reraise_exception(): [ 794.267200] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] self.force_reraise() [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] raise self.value [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] updated_port = self._update_port( [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] _ensure_no_port_binding_failure(port) [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] raise exception.PortBindingFailed(port_id=port['id']) [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] nova.exception.PortBindingFailed: Binding failed for port 260b6560-883f-441f-ada4-45fef2d173d6, please check neutron logs for more information. [ 794.267641] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] [ 794.267641] env[63355]: INFO nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Terminating instance [ 794.268756] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Acquiring lock "refresh_cache-282f1b55-8a74-467a-9ec8-0684eb780970" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.396051] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.396051] env[63355]: ERROR nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 407f76f2-db4c-42a7-a5e7-68033196ad71, please check neutron logs for more information. [ 794.396051] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Traceback (most recent call last): [ 794.396051] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 794.396051] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] self.driver.spawn(context, instance, image_meta, [ 794.396051] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 794.396051] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 794.396051] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 794.396051] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] vm_ref = self.build_virtual_machine(instance, [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] for vif in network_info: [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] return self._sync_wrapper(fn, *args, **kwargs) [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] self.wait() [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] self[:] = self._gt.wait() [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] return self._exit_event.wait() [ 794.396571] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] result = hub.switch() [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] return self.greenlet.switch() [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] result = function(*args, **kwargs) [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] return func(*args, **kwargs) [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] raise e [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] nwinfo = self.network_api.allocate_for_instance( [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 794.397239] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] created_port_ids = self._update_ports_for_instance( [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] with excutils.save_and_reraise_exception(): [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] self.force_reraise() [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] raise self.value [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] updated_port = self._update_port( [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] _ensure_no_port_binding_failure(port) [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.397837] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] raise exception.PortBindingFailed(port_id=port['id']) [ 794.398386] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] nova.exception.PortBindingFailed: Binding failed for port 407f76f2-db4c-42a7-a5e7-68033196ad71, please check neutron logs for more information. [ 794.398386] env[63355]: ERROR nova.compute.manager [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] [ 794.398386] env[63355]: DEBUG nova.compute.utils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Binding failed for port 407f76f2-db4c-42a7-a5e7-68033196ad71, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 794.398386] env[63355]: DEBUG nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Build of instance 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6 was re-scheduled: Binding failed for port 407f76f2-db4c-42a7-a5e7-68033196ad71, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 794.398601] env[63355]: DEBUG nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 794.398815] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Acquiring lock "refresh_cache-8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.398959] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Acquired lock "refresh_cache-8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.399130] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.400110] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.848s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.653554] env[63355]: DEBUG oslo_concurrency.lockutils [req-8f037691-6c4d-4407-bfb6-265abba14719 req-3f8021fb-5fc8-46d3-8cd1-25d0c9e1838d service nova] Releasing lock "refresh_cache-282f1b55-8a74-467a-9ec8-0684eb780970" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.653886] env[63355]: DEBUG nova.compute.manager [req-8f037691-6c4d-4407-bfb6-265abba14719 req-3f8021fb-5fc8-46d3-8cd1-25d0c9e1838d service nova] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Received event network-vif-deleted-260b6560-883f-441f-ada4-45fef2d173d6 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.654287] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Acquired lock "refresh_cache-282f1b55-8a74-467a-9ec8-0684eb780970" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.654460] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.730736] env[63355]: INFO nova.compute.manager [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] [instance: d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6] Took 1.15 seconds to deallocate network for instance. [ 794.923109] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.995109] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.175629] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.250688] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4841b6-306a-4944-87cd-3b867874773e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.253950] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.261579] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290d06db-fe84-4368-96ff-6fb21d9475ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.292033] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9cc4de-4fac-429e-b327-e5d072f32457 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.300177] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381a082b-714a-4d61-abc2-602248f1c874 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.313675] env[63355]: DEBUG nova.compute.provider_tree [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.499104] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Releasing lock "refresh_cache-8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.499317] env[63355]: DEBUG nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 795.499498] env[63355]: DEBUG nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 795.499662] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 795.514944] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.757757] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Releasing lock "refresh_cache-282f1b55-8a74-467a-9ec8-0684eb780970" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.758505] env[63355]: DEBUG nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 795.758682] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae60a6a1-df2c-4e8e-9108-32270b266540 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.761773] env[63355]: INFO nova.scheduler.client.report [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Deleted allocations for instance d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6 [ 795.773564] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6317dfb0-3a13-4797-b0a0-d13562d810c8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.795943] env[63355]: WARNING nova.virt.vmwareapi.driver [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 282f1b55-8a74-467a-9ec8-0684eb780970 could not be found. [ 795.796678] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 795.796678] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4224bd84-41de-408e-94b5-cead34fa3b7d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.803907] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f01b3b-5061-4d64-85df-c85be044d2b9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.815926] env[63355]: DEBUG nova.scheduler.client.report [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.832429] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 282f1b55-8a74-467a-9ec8-0684eb780970 could not be found. [ 795.832732] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 795.832952] env[63355]: INFO nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Took 0.07 seconds to destroy the instance on the hypervisor. [ 795.833274] env[63355]: DEBUG oslo.service.loopingcall [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.833552] env[63355]: DEBUG nova.compute.manager [-] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 795.833712] env[63355]: DEBUG nova.network.neutron [-] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 795.848592] env[63355]: DEBUG nova.network.neutron [-] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.017903] env[63355]: DEBUG nova.network.neutron [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.269186] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69557818-d1e4-4cf4-a5aa-0ebeec038117 tempest-ListServerFiltersTestJSON-1551439793 tempest-ListServerFiltersTestJSON-1551439793-project-member] Lock "d5daa7b5-2b00-46f5-b89b-42f33f6cd3c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.891s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.322046] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.921s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.322357] env[63355]: ERROR nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fb7b1d5f-c8fa-4ae5-8522-75ae880ae067, please check neutron logs for more information. [ 796.322357] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Traceback (most recent call last): [ 796.322357] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 796.322357] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] self.driver.spawn(context, instance, image_meta, [ 796.322357] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 796.322357] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 796.322357] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 796.322357] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] vm_ref = self.build_virtual_machine(instance, [ 796.322357] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 796.322357] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 796.322357] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] for vif in network_info: [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] return self._sync_wrapper(fn, *args, **kwargs) [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] self.wait() [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] self[:] = self._gt.wait() [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] return self._exit_event.wait() [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] current.throw(*self._exc) [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 796.322760] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] result = function(*args, **kwargs) [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] return func(*args, **kwargs) [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] raise e [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] nwinfo = self.network_api.allocate_for_instance( [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] created_port_ids = self._update_ports_for_instance( [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] with excutils.save_and_reraise_exception(): [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] self.force_reraise() [ 796.323212] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.323636] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] raise self.value [ 796.323636] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 796.323636] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] updated_port = self._update_port( [ 796.323636] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.323636] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] _ensure_no_port_binding_failure(port) [ 796.323636] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.323636] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] raise exception.PortBindingFailed(port_id=port['id']) [ 796.323636] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] nova.exception.PortBindingFailed: Binding failed for port fb7b1d5f-c8fa-4ae5-8522-75ae880ae067, please check neutron logs for more information. [ 796.323636] env[63355]: ERROR nova.compute.manager [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] [ 796.323636] env[63355]: DEBUG nova.compute.utils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Binding failed for port fb7b1d5f-c8fa-4ae5-8522-75ae880ae067, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 796.324583] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.273s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.326266] env[63355]: INFO nova.compute.claims [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.329265] env[63355]: DEBUG nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Build of instance 1446c7da-415f-43ee-9ff1-2266bd3038e4 was re-scheduled: Binding failed for port fb7b1d5f-c8fa-4ae5-8522-75ae880ae067, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 796.329686] env[63355]: DEBUG nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 796.330106] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "refresh_cache-1446c7da-415f-43ee-9ff1-2266bd3038e4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.330267] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquired lock "refresh_cache-1446c7da-415f-43ee-9ff1-2266bd3038e4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.330429] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 796.353565] env[63355]: DEBUG nova.network.neutron [-] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.520216] env[63355]: INFO nova.compute.manager [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] [instance: 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6] Took 1.02 seconds to deallocate network for instance. [ 796.773024] env[63355]: DEBUG nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.849386] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.855954] env[63355]: INFO nova.compute.manager [-] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Took 1.02 seconds to deallocate network for instance. [ 796.938506] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.310153] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.440411] env[63355]: INFO nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Took 0.58 seconds to detach 1 volumes for instance. [ 797.442447] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Releasing lock "refresh_cache-1446c7da-415f-43ee-9ff1-2266bd3038e4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.442447] env[63355]: DEBUG nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 797.442447] env[63355]: DEBUG nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.442672] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 797.444288] env[63355]: DEBUG nova.compute.claims [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 797.444454] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.459884] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.547111] env[63355]: INFO nova.scheduler.client.report [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Deleted allocations for instance 8e5e4a26-69ed-461d-aa67-929c0a4ab0f6 [ 797.731731] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd61fe9-6be1-48b0-864a-9afb40957204 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.739265] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf29996-4ffb-4560-aa58-62770dc6aa37 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.768462] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0526742-9721-4a3b-af92-d4f2b1b4fcae {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.775722] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a47379-5542-4d26-bf68-27f1124f723f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.788472] env[63355]: DEBUG nova.compute.provider_tree [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.963328] env[63355]: DEBUG nova.network.neutron [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.059887] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cebb559a-6ce5-4a6e-8af4-40105b3edc35 tempest-ServersTestFqdnHostnames-1651411780 tempest-ServersTestFqdnHostnames-1651411780-project-member] Lock "8e5e4a26-69ed-461d-aa67-929c0a4ab0f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.615s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.292163] env[63355]: DEBUG nova.scheduler.client.report [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.466711] env[63355]: INFO nova.compute.manager [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 1446c7da-415f-43ee-9ff1-2266bd3038e4] Took 1.02 seconds to deallocate network for instance. [ 798.562545] env[63355]: DEBUG nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 798.798142] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.798687] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 798.801863] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.438s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.088637] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.303346] env[63355]: DEBUG nova.compute.utils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.304967] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 799.305152] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 799.352698] env[63355]: DEBUG nova.policy [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4382845ae7694c018aa5c0562316583a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '092fc5912cd4449895f1c4d1318dd390', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 799.493371] env[63355]: INFO nova.scheduler.client.report [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Deleted allocations for instance 1446c7da-415f-43ee-9ff1-2266bd3038e4 [ 799.632534] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Successfully created port: 8b6ff543-f0de-4659-8c3f-93517f56fc21 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.713645] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f40d06d-c65e-464c-9fec-c5926b166709 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.721752] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a1a058-0b4a-477b-ba3c-9610fe381e3b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.751999] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befe48c1-f8ba-42b9-90ee-8be5ebb45d5f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.759864] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b767dbae-5aa4-4878-a38f-de2c1aacacc3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.773296] env[63355]: DEBUG nova.compute.provider_tree [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.815535] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.004386] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9e908101-e4d9-4aa7-9569-59cc1977cfc8 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "1446c7da-415f-43ee-9ff1-2266bd3038e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 160.785s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.281334] env[63355]: DEBUG nova.scheduler.client.report [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.346335] env[63355]: DEBUG nova.compute.manager [req-c5b0dde9-8c65-489b-a7be-4a0d17c46f7e req-1243c65c-f4b9-4c21-b3aa-221b0b7583e8 service nova] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Received event network-changed-8b6ff543-f0de-4659-8c3f-93517f56fc21 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.346550] env[63355]: DEBUG nova.compute.manager [req-c5b0dde9-8c65-489b-a7be-4a0d17c46f7e req-1243c65c-f4b9-4c21-b3aa-221b0b7583e8 service nova] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Refreshing instance network info cache due to event network-changed-8b6ff543-f0de-4659-8c3f-93517f56fc21. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 800.346764] env[63355]: DEBUG oslo_concurrency.lockutils [req-c5b0dde9-8c65-489b-a7be-4a0d17c46f7e req-1243c65c-f4b9-4c21-b3aa-221b0b7583e8 service nova] Acquiring lock "refresh_cache-7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.346907] env[63355]: DEBUG oslo_concurrency.lockutils [req-c5b0dde9-8c65-489b-a7be-4a0d17c46f7e req-1243c65c-f4b9-4c21-b3aa-221b0b7583e8 service nova] Acquired lock "refresh_cache-7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.347294] env[63355]: DEBUG nova.network.neutron [req-c5b0dde9-8c65-489b-a7be-4a0d17c46f7e req-1243c65c-f4b9-4c21-b3aa-221b0b7583e8 service nova] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Refreshing network info cache for port 8b6ff543-f0de-4659-8c3f-93517f56fc21 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.501370] env[63355]: ERROR nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8b6ff543-f0de-4659-8c3f-93517f56fc21, please check neutron logs for more information. [ 800.501370] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 800.501370] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.501370] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 800.501370] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.501370] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 800.501370] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.501370] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 800.501370] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.501370] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 800.501370] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.501370] env[63355]: ERROR nova.compute.manager raise self.value [ 800.501370] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.501370] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 800.501370] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.501370] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 800.501929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.501929] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 800.501929] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8b6ff543-f0de-4659-8c3f-93517f56fc21, please check neutron logs for more information. [ 800.501929] env[63355]: ERROR nova.compute.manager [ 800.501929] env[63355]: Traceback (most recent call last): [ 800.501929] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 800.501929] env[63355]: listener.cb(fileno) [ 800.501929] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.501929] env[63355]: result = function(*args, **kwargs) [ 800.501929] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 800.501929] env[63355]: return func(*args, **kwargs) [ 800.501929] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.501929] env[63355]: raise e [ 800.501929] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.501929] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 800.501929] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.501929] env[63355]: created_port_ids = self._update_ports_for_instance( [ 800.501929] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.501929] env[63355]: with excutils.save_and_reraise_exception(): [ 800.501929] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.501929] env[63355]: self.force_reraise() [ 800.501929] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.501929] env[63355]: raise self.value [ 800.501929] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.501929] env[63355]: updated_port = self._update_port( [ 800.501929] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.501929] env[63355]: _ensure_no_port_binding_failure(port) [ 800.501929] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.501929] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 800.503026] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 8b6ff543-f0de-4659-8c3f-93517f56fc21, please check neutron logs for more information. [ 800.503026] env[63355]: Removing descriptor: 17 [ 800.507147] env[63355]: DEBUG nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.786800] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.985s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.787484] env[63355]: ERROR nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 15746d33-dfc2-452b-a522-8cf17fbec8c0, please check neutron logs for more information. [ 800.787484] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Traceback (most recent call last): [ 800.787484] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 800.787484] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] self.driver.spawn(context, instance, image_meta, [ 800.787484] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 800.787484] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 800.787484] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 800.787484] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] vm_ref = self.build_virtual_machine(instance, [ 800.787484] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 800.787484] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 800.787484] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] for vif in network_info: [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] return self._sync_wrapper(fn, *args, **kwargs) [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] self.wait() [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] self[:] = self._gt.wait() [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] return self._exit_event.wait() [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] result = hub.switch() [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 800.787856] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] return self.greenlet.switch() [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] result = function(*args, **kwargs) [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] return func(*args, **kwargs) [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] raise e [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] nwinfo = self.network_api.allocate_for_instance( [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] created_port_ids = self._update_ports_for_instance( [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] with excutils.save_and_reraise_exception(): [ 800.788252] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] self.force_reraise() [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] raise self.value [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] updated_port = self._update_port( [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] _ensure_no_port_binding_failure(port) [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] raise exception.PortBindingFailed(port_id=port['id']) [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] nova.exception.PortBindingFailed: Binding failed for port 15746d33-dfc2-452b-a522-8cf17fbec8c0, please check neutron logs for more information. [ 800.788622] env[63355]: ERROR nova.compute.manager [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] [ 800.788940] env[63355]: DEBUG nova.compute.utils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Binding failed for port 15746d33-dfc2-452b-a522-8cf17fbec8c0, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 800.789799] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.168s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.791276] env[63355]: INFO nova.compute.claims [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.793827] env[63355]: DEBUG nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Build of instance ec4726b2-38f9-4499-a5b0-a70db0d697f6 was re-scheduled: Binding failed for port 15746d33-dfc2-452b-a522-8cf17fbec8c0, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 800.794260] env[63355]: DEBUG nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 800.794474] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Acquiring lock "refresh_cache-ec4726b2-38f9-4499-a5b0-a70db0d697f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.794723] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Acquired lock "refresh_cache-ec4726b2-38f9-4499-a5b0-a70db0d697f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.794921] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 800.828585] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 800.860661] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.860912] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.861076] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.861260] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.861404] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.861547] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.861775] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.863562] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.863562] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.863562] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.863562] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.863562] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba5bf360-358a-486a-9c22-21d1d3ccfc10 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.872714] env[63355]: DEBUG nova.network.neutron [req-c5b0dde9-8c65-489b-a7be-4a0d17c46f7e req-1243c65c-f4b9-4c21-b3aa-221b0b7583e8 service nova] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.875260] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f5bb76-b3d3-4716-bd46-7cb7a6998081 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.893018] env[63355]: ERROR nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8b6ff543-f0de-4659-8c3f-93517f56fc21, please check neutron logs for more information. [ 800.893018] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Traceback (most recent call last): [ 800.893018] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 800.893018] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] yield resources [ 800.893018] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 800.893018] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] self.driver.spawn(context, instance, image_meta, [ 800.893018] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 800.893018] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 800.893018] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 800.893018] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] vm_ref = self.build_virtual_machine(instance, [ 800.893018] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] for vif in network_info: [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] return self._sync_wrapper(fn, *args, **kwargs) [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] self.wait() [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] self[:] = self._gt.wait() [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] return self._exit_event.wait() [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 800.893698] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] current.throw(*self._exc) [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] result = function(*args, **kwargs) [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] return func(*args, **kwargs) [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] raise e [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] nwinfo = self.network_api.allocate_for_instance( [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] created_port_ids = self._update_ports_for_instance( [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] with excutils.save_and_reraise_exception(): [ 800.894208] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] self.force_reraise() [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] raise self.value [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] updated_port = self._update_port( [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] _ensure_no_port_binding_failure(port) [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] raise exception.PortBindingFailed(port_id=port['id']) [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] nova.exception.PortBindingFailed: Binding failed for port 8b6ff543-f0de-4659-8c3f-93517f56fc21, please check neutron logs for more information. [ 800.894617] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] [ 800.894617] env[63355]: INFO nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Terminating instance [ 800.895161] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "refresh_cache-7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.026034] env[63355]: DEBUG nova.network.neutron [req-c5b0dde9-8c65-489b-a7be-4a0d17c46f7e req-1243c65c-f4b9-4c21-b3aa-221b0b7583e8 service nova] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.031774] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.317359] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.393236] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.529246] env[63355]: DEBUG oslo_concurrency.lockutils [req-c5b0dde9-8c65-489b-a7be-4a0d17c46f7e req-1243c65c-f4b9-4c21-b3aa-221b0b7583e8 service nova] Releasing lock "refresh_cache-7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.529659] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired lock "refresh_cache-7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.529866] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.896296] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Releasing lock "refresh_cache-ec4726b2-38f9-4499-a5b0-a70db0d697f6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.896581] env[63355]: DEBUG nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 801.896794] env[63355]: DEBUG nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 801.896872] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 801.918442] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.051093] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.149040] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.203048] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e7a4eb-20e5-4171-a353-06b102873008 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.210554] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62665bb8-6c37-45c4-8f54-7629e588a0cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.240250] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df24c01-6288-4403-9734-971a4cfd8456 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.247694] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad67b80a-bbda-4eed-a1c9-8d136d408f69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.260975] env[63355]: DEBUG nova.compute.provider_tree [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.421470] env[63355]: DEBUG nova.network.neutron [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.499324] env[63355]: DEBUG nova.compute.manager [req-fb5a29f0-13be-4e04-948c-cf823b6be02c req-c74fbd92-1d3f-4a47-a9b4-c1972d18282e service nova] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Received event network-vif-deleted-8b6ff543-f0de-4659-8c3f-93517f56fc21 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.651649] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Releasing lock "refresh_cache-7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.652210] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 802.652456] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 802.652766] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-308f5992-3804-45b3-8217-6abf0bea96e8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.661518] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f3b1fc-dcb8-4a10-a26e-508f1f596035 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.685864] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d could not be found. [ 802.686488] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 802.686488] env[63355]: INFO nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 802.686638] env[63355]: DEBUG oslo.service.loopingcall [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.686851] env[63355]: DEBUG nova.compute.manager [-] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 802.686935] env[63355]: DEBUG nova.network.neutron [-] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 802.705573] env[63355]: DEBUG nova.network.neutron [-] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.767159] env[63355]: DEBUG nova.scheduler.client.report [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.927729] env[63355]: INFO nova.compute.manager [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] [instance: ec4726b2-38f9-4499-a5b0-a70db0d697f6] Took 1.03 seconds to deallocate network for instance. [ 803.207645] env[63355]: DEBUG nova.network.neutron [-] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.274217] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.274217] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.276369] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.713s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.277860] env[63355]: INFO nova.compute.claims [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.713450] env[63355]: INFO nova.compute.manager [-] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Took 1.03 seconds to deallocate network for instance. [ 803.718197] env[63355]: DEBUG nova.compute.claims [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 803.718400] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.782908] env[63355]: DEBUG nova.compute.utils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.787507] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.787716] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.832098] env[63355]: DEBUG nova.policy [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4382845ae7694c018aa5c0562316583a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '092fc5912cd4449895f1c4d1318dd390', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 804.010440] env[63355]: INFO nova.scheduler.client.report [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Deleted allocations for instance ec4726b2-38f9-4499-a5b0-a70db0d697f6 [ 804.277190] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "74071e73-10fa-4dcb-aa15-91303b2278fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.277439] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "74071e73-10fa-4dcb-aa15-91303b2278fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.289598] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.300198] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Successfully created port: af5b9176-6396-4c83-8612-ed8c639cbcd4 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.524270] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a804a09e-69de-41e8-bc85-fc661a4d64f6 tempest-ServersTestJSON-1828804353 tempest-ServersTestJSON-1828804353-project-member] Lock "ec4726b2-38f9-4499-a5b0-a70db0d697f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.951s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.735065] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e9fcaf-b41d-45ff-948f-c5094d6a089b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.742866] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b8d279-c1f7-42ea-b725-fab99f0022c8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.772203] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e17ea0-bc37-4b53-9651-f08148b8c1d6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.779742] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abf67494-3e7c-4ae6-8288-863999ed3404 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.793919] env[63355]: DEBUG nova.compute.provider_tree [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.026986] env[63355]: DEBUG nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.297043] env[63355]: DEBUG nova.scheduler.client.report [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.302666] env[63355]: DEBUG nova.compute.manager [req-33a032c6-86e1-4d16-9909-fe1b8d546a3c req-e86a1c72-aa00-49b7-863f-8c8f2b51fdb7 service nova] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Received event network-changed-af5b9176-6396-4c83-8612-ed8c639cbcd4 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.302894] env[63355]: DEBUG nova.compute.manager [req-33a032c6-86e1-4d16-9909-fe1b8d546a3c req-e86a1c72-aa00-49b7-863f-8c8f2b51fdb7 service nova] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Refreshing instance network info cache due to event network-changed-af5b9176-6396-4c83-8612-ed8c639cbcd4. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 805.303146] env[63355]: DEBUG oslo_concurrency.lockutils [req-33a032c6-86e1-4d16-9909-fe1b8d546a3c req-e86a1c72-aa00-49b7-863f-8c8f2b51fdb7 service nova] Acquiring lock "refresh_cache-795b61f6-a0f3-4aab-94b3-2540ccbb7312" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.303473] env[63355]: DEBUG oslo_concurrency.lockutils [req-33a032c6-86e1-4d16-9909-fe1b8d546a3c req-e86a1c72-aa00-49b7-863f-8c8f2b51fdb7 service nova] Acquired lock "refresh_cache-795b61f6-a0f3-4aab-94b3-2540ccbb7312" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.303473] env[63355]: DEBUG nova.network.neutron [req-33a032c6-86e1-4d16-9909-fe1b8d546a3c req-e86a1c72-aa00-49b7-863f-8c8f2b51fdb7 service nova] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Refreshing network info cache for port af5b9176-6396-4c83-8612-ed8c639cbcd4 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.305919] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.342023] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.342023] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.342023] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.342272] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.342272] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.342272] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.342272] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.349665] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.349665] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.349665] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.349665] env[63355]: DEBUG nova.virt.hardware [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.349665] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c24f90-e3c2-44b6-a6d9-b3d56f1ebb99 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.361998] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d869c0c-d2dc-4dbe-ad5b-75f6d544f660 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.412199] env[63355]: ERROR nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port af5b9176-6396-4c83-8612-ed8c639cbcd4, please check neutron logs for more information. [ 805.412199] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 805.412199] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 805.412199] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 805.412199] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 805.412199] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 805.412199] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 805.412199] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 805.412199] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.412199] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 805.412199] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.412199] env[63355]: ERROR nova.compute.manager raise self.value [ 805.412199] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 805.412199] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 805.412199] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.412199] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 805.412853] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.412853] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 805.412853] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port af5b9176-6396-4c83-8612-ed8c639cbcd4, please check neutron logs for more information. [ 805.412853] env[63355]: ERROR nova.compute.manager [ 805.412853] env[63355]: Traceback (most recent call last): [ 805.412853] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 805.412853] env[63355]: listener.cb(fileno) [ 805.412853] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 805.412853] env[63355]: result = function(*args, **kwargs) [ 805.412853] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 805.412853] env[63355]: return func(*args, **kwargs) [ 805.412853] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 805.412853] env[63355]: raise e [ 805.412853] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 805.412853] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 805.412853] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 805.412853] env[63355]: created_port_ids = self._update_ports_for_instance( [ 805.412853] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 805.412853] env[63355]: with excutils.save_and_reraise_exception(): [ 805.412853] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.412853] env[63355]: self.force_reraise() [ 805.412853] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.412853] env[63355]: raise self.value [ 805.412853] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 805.412853] env[63355]: updated_port = self._update_port( [ 805.412853] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.412853] env[63355]: _ensure_no_port_binding_failure(port) [ 805.412853] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.412853] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 805.414654] env[63355]: nova.exception.PortBindingFailed: Binding failed for port af5b9176-6396-4c83-8612-ed8c639cbcd4, please check neutron logs for more information. [ 805.414654] env[63355]: Removing descriptor: 17 [ 805.414654] env[63355]: ERROR nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port af5b9176-6396-4c83-8612-ed8c639cbcd4, please check neutron logs for more information. [ 805.414654] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Traceback (most recent call last): [ 805.414654] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 805.414654] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] yield resources [ 805.414654] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 805.414654] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] self.driver.spawn(context, instance, image_meta, [ 805.414654] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 805.414654] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] self._vmops.spawn(context, instance, image_meta, injected_files, [ 805.414654] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 805.414654] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] vm_ref = self.build_virtual_machine(instance, [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] vif_infos = vmwarevif.get_vif_info(self._session, [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] for vif in network_info: [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] return self._sync_wrapper(fn, *args, **kwargs) [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] self.wait() [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] self[:] = self._gt.wait() [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] return self._exit_event.wait() [ 805.415663] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] result = hub.switch() [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] return self.greenlet.switch() [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] result = function(*args, **kwargs) [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] return func(*args, **kwargs) [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] raise e [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] nwinfo = self.network_api.allocate_for_instance( [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 805.416594] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] created_port_ids = self._update_ports_for_instance( [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] with excutils.save_and_reraise_exception(): [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] self.force_reraise() [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] raise self.value [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] updated_port = self._update_port( [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] _ensure_no_port_binding_failure(port) [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.417807] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] raise exception.PortBindingFailed(port_id=port['id']) [ 805.418388] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] nova.exception.PortBindingFailed: Binding failed for port af5b9176-6396-4c83-8612-ed8c639cbcd4, please check neutron logs for more information. [ 805.418388] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] [ 805.418388] env[63355]: INFO nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Terminating instance [ 805.418388] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "refresh_cache-795b61f6-a0f3-4aab-94b3-2540ccbb7312" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.549362] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.808118] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.808840] env[63355]: DEBUG nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 805.817810] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.187s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.842172] env[63355]: DEBUG nova.network.neutron [req-33a032c6-86e1-4d16-9909-fe1b8d546a3c req-e86a1c72-aa00-49b7-863f-8c8f2b51fdb7 service nova] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.939686] env[63355]: DEBUG nova.network.neutron [req-33a032c6-86e1-4d16-9909-fe1b8d546a3c req-e86a1c72-aa00-49b7-863f-8c8f2b51fdb7 service nova] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.215893] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 806.215893] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 806.325882] env[63355]: DEBUG nova.compute.utils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 806.332062] env[63355]: DEBUG nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 806.332062] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 806.399599] env[63355]: DEBUG nova.policy [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a170b12d59274e4dbebc8a7c73b603a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '838c804beaeb49a3b0f84af63529b6c9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 806.442997] env[63355]: DEBUG oslo_concurrency.lockutils [req-33a032c6-86e1-4d16-9909-fe1b8d546a3c req-e86a1c72-aa00-49b7-863f-8c8f2b51fdb7 service nova] Releasing lock "refresh_cache-795b61f6-a0f3-4aab-94b3-2540ccbb7312" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.443431] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired lock "refresh_cache-795b61f6-a0f3-4aab-94b3-2540ccbb7312" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.443617] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.724465] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 806.724465] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 806.724465] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Rebuilding the list of instances to heal {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 806.746587] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d4a882-752b-43b9-8069-095e377174ae {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.756542] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a57b94-3a72-4aeb-b825-614959a343fa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.789525] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dcff87f-1123-479b-a697-3f1318c12d64 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.796745] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96255a37-3d5d-4d51-b1ad-ebc153be43b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.812054] env[63355]: DEBUG nova.compute.provider_tree [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.832310] env[63355]: DEBUG nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 806.895367] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Successfully created port: f4801f4b-2b95-41b6-a74c-7c95d83c7152 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.976502] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.027989] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.028249] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.176676] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.231914] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 807.232194] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 807.232287] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 807.233087] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 807.233087] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 807.233087] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Didn't find any instances for network info cache update. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 807.233087] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 807.233087] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 807.233294] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 807.233294] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 807.233383] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 807.233520] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 807.233640] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 807.233773] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 807.316638] env[63355]: DEBUG nova.scheduler.client.report [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.345440] env[63355]: DEBUG nova.compute.manager [req-0027ecbd-8964-446f-aa03-09427c947520 req-fc2a902a-f98b-4ed0-ad37-3c68f803e618 service nova] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Received event network-vif-deleted-af5b9176-6396-4c83-8612-ed8c639cbcd4 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.679557] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Releasing lock "refresh_cache-795b61f6-a0f3-4aab-94b3-2540ccbb7312" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.680073] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.680282] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.680590] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c8a37cf-2fd9-4223-8c73-8622b4ce7f9c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.690228] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abcabd6-d557-4816-9933-e0e09482dde3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.712314] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 795b61f6-a0f3-4aab-94b3-2540ccbb7312 could not be found. [ 807.712533] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 807.712781] env[63355]: INFO nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Took 0.03 seconds to destroy the instance on the hypervisor. [ 807.713725] env[63355]: DEBUG oslo.service.loopingcall [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.713725] env[63355]: DEBUG nova.compute.manager [-] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 807.713725] env[63355]: DEBUG nova.network.neutron [-] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 807.731240] env[63355]: DEBUG nova.network.neutron [-] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.739579] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.821398] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.006s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.821631] env[63355]: ERROR nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 34d3c5d5-253d-4236-9f68-078074585482, please check neutron logs for more information. [ 807.821631] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Traceback (most recent call last): [ 807.821631] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 807.821631] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] self.driver.spawn(context, instance, image_meta, [ 807.821631] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 807.821631] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] self._vmops.spawn(context, instance, image_meta, injected_files, [ 807.821631] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 807.821631] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] vm_ref = self.build_virtual_machine(instance, [ 807.821631] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 807.821631] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] vif_infos = vmwarevif.get_vif_info(self._session, [ 807.821631] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] for vif in network_info: [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] return self._sync_wrapper(fn, *args, **kwargs) [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] self.wait() [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] self[:] = self._gt.wait() [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] return self._exit_event.wait() [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] result = hub.switch() [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 807.822012] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] return self.greenlet.switch() [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] result = function(*args, **kwargs) [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] return func(*args, **kwargs) [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] raise e [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] nwinfo = self.network_api.allocate_for_instance( [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] created_port_ids = self._update_ports_for_instance( [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] with excutils.save_and_reraise_exception(): [ 807.822479] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] self.force_reraise() [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] raise self.value [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] updated_port = self._update_port( [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] _ensure_no_port_binding_failure(port) [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] raise exception.PortBindingFailed(port_id=port['id']) [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] nova.exception.PortBindingFailed: Binding failed for port 34d3c5d5-253d-4236-9f68-078074585482, please check neutron logs for more information. [ 807.822906] env[63355]: ERROR nova.compute.manager [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] [ 807.823268] env[63355]: DEBUG nova.compute.utils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Binding failed for port 34d3c5d5-253d-4236-9f68-078074585482, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 807.823558] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.335s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.828321] env[63355]: DEBUG nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Build of instance 3c04c546-a9fa-4cbd-98d4-3ea92918a918 was re-scheduled: Binding failed for port 34d3c5d5-253d-4236-9f68-078074585482, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 807.828826] env[63355]: DEBUG nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 807.829135] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Acquiring lock "refresh_cache-3c04c546-a9fa-4cbd-98d4-3ea92918a918" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.829203] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Acquired lock "refresh_cache-3c04c546-a9fa-4cbd-98d4-3ea92918a918" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.829356] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.847434] env[63355]: DEBUG nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 807.886792] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 807.886792] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 807.886977] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.887034] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 807.887151] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.887289] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 807.887486] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 807.887636] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 807.887795] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 807.887999] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 807.888181] env[63355]: DEBUG nova.virt.hardware [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.889283] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09385f3-211b-4976-b00b-d182ed5f09fc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.897266] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad91eee-0e8c-4edb-a6cc-c7d988fb2c00 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.052027] env[63355]: ERROR nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f4801f4b-2b95-41b6-a74c-7c95d83c7152, please check neutron logs for more information. [ 808.052027] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 808.052027] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 808.052027] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 808.052027] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 808.052027] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 808.052027] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 808.052027] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 808.052027] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.052027] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 808.052027] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.052027] env[63355]: ERROR nova.compute.manager raise self.value [ 808.052027] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 808.052027] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 808.052027] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 808.052027] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 808.052695] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 808.052695] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 808.052695] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f4801f4b-2b95-41b6-a74c-7c95d83c7152, please check neutron logs for more information. [ 808.052695] env[63355]: ERROR nova.compute.manager [ 808.052695] env[63355]: Traceback (most recent call last): [ 808.052695] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 808.052695] env[63355]: listener.cb(fileno) [ 808.052695] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 808.052695] env[63355]: result = function(*args, **kwargs) [ 808.052695] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 808.052695] env[63355]: return func(*args, **kwargs) [ 808.052695] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 808.052695] env[63355]: raise e [ 808.052695] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 808.052695] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 808.052695] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 808.052695] env[63355]: created_port_ids = self._update_ports_for_instance( [ 808.052695] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 808.052695] env[63355]: with excutils.save_and_reraise_exception(): [ 808.052695] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.052695] env[63355]: self.force_reraise() [ 808.052695] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.052695] env[63355]: raise self.value [ 808.052695] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 808.052695] env[63355]: updated_port = self._update_port( [ 808.052695] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 808.052695] env[63355]: _ensure_no_port_binding_failure(port) [ 808.052695] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 808.052695] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 808.053664] env[63355]: nova.exception.PortBindingFailed: Binding failed for port f4801f4b-2b95-41b6-a74c-7c95d83c7152, please check neutron logs for more information. [ 808.053664] env[63355]: Removing descriptor: 17 [ 808.053664] env[63355]: ERROR nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f4801f4b-2b95-41b6-a74c-7c95d83c7152, please check neutron logs for more information. [ 808.053664] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Traceback (most recent call last): [ 808.053664] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 808.053664] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] yield resources [ 808.053664] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 808.053664] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] self.driver.spawn(context, instance, image_meta, [ 808.053664] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 808.053664] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 808.053664] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 808.053664] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] vm_ref = self.build_virtual_machine(instance, [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] for vif in network_info: [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] return self._sync_wrapper(fn, *args, **kwargs) [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] self.wait() [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] self[:] = self._gt.wait() [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] return self._exit_event.wait() [ 808.054259] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] result = hub.switch() [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] return self.greenlet.switch() [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] result = function(*args, **kwargs) [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] return func(*args, **kwargs) [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] raise e [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] nwinfo = self.network_api.allocate_for_instance( [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 808.055010] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] created_port_ids = self._update_ports_for_instance( [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] with excutils.save_and_reraise_exception(): [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] self.force_reraise() [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] raise self.value [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] updated_port = self._update_port( [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] _ensure_no_port_binding_failure(port) [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 808.055468] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] raise exception.PortBindingFailed(port_id=port['id']) [ 808.055828] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] nova.exception.PortBindingFailed: Binding failed for port f4801f4b-2b95-41b6-a74c-7c95d83c7152, please check neutron logs for more information. [ 808.055828] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] [ 808.055828] env[63355]: INFO nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Terminating instance [ 808.055828] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Acquiring lock "refresh_cache-7816d808-c9dd-403f-b993-c6d9dc458c7b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.055828] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Acquired lock "refresh_cache-7816d808-c9dd-403f-b993-c6d9dc458c7b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.055828] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.233700] env[63355]: DEBUG nova.network.neutron [-] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.353373] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.439338] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.578912] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.658442] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.719769] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951bea26-35c8-4582-ba46-859a6d591b51 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.727792] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17a82ad-06f2-4d23-9dcf-4b0e9c2d5b8b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.761035] env[63355]: INFO nova.compute.manager [-] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Took 1.05 seconds to deallocate network for instance. [ 808.762424] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae68235-cfca-428b-aebd-6ff78275dccc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.765243] env[63355]: DEBUG nova.compute.claims [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 808.765418] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.770510] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da34092-700f-403b-a641-61f2f380b727 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.788132] env[63355]: DEBUG nova.compute.provider_tree [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.941792] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Releasing lock "refresh_cache-3c04c546-a9fa-4cbd-98d4-3ea92918a918" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.941928] env[63355]: DEBUG nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 808.942078] env[63355]: DEBUG nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.942250] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.959433] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.161020] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Releasing lock "refresh_cache-7816d808-c9dd-403f-b993-c6d9dc458c7b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.161344] env[63355]: DEBUG nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 809.161962] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 809.161962] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a0da0af-b328-4e10-9044-5a2671f9dda9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.170704] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba04948-5b90-4006-849c-70902fa1c4ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.191526] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7816d808-c9dd-403f-b993-c6d9dc458c7b could not be found. [ 809.191738] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 809.191916] env[63355]: INFO nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 809.192170] env[63355]: DEBUG oslo.service.loopingcall [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.192382] env[63355]: DEBUG nova.compute.manager [-] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.192474] env[63355]: DEBUG nova.network.neutron [-] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.206994] env[63355]: DEBUG nova.network.neutron [-] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.291221] env[63355]: DEBUG nova.scheduler.client.report [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.363955] env[63355]: DEBUG nova.compute.manager [req-6a856bd6-e970-408f-aa27-e5d585d06a8a req-ae20baf5-0cc5-4d3f-8be9-3f23dd796dca service nova] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Received event network-changed-f4801f4b-2b95-41b6-a74c-7c95d83c7152 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.364125] env[63355]: DEBUG nova.compute.manager [req-6a856bd6-e970-408f-aa27-e5d585d06a8a req-ae20baf5-0cc5-4d3f-8be9-3f23dd796dca service nova] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Refreshing instance network info cache due to event network-changed-f4801f4b-2b95-41b6-a74c-7c95d83c7152. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 809.364338] env[63355]: DEBUG oslo_concurrency.lockutils [req-6a856bd6-e970-408f-aa27-e5d585d06a8a req-ae20baf5-0cc5-4d3f-8be9-3f23dd796dca service nova] Acquiring lock "refresh_cache-7816d808-c9dd-403f-b993-c6d9dc458c7b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.364477] env[63355]: DEBUG oslo_concurrency.lockutils [req-6a856bd6-e970-408f-aa27-e5d585d06a8a req-ae20baf5-0cc5-4d3f-8be9-3f23dd796dca service nova] Acquired lock "refresh_cache-7816d808-c9dd-403f-b993-c6d9dc458c7b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.364634] env[63355]: DEBUG nova.network.neutron [req-6a856bd6-e970-408f-aa27-e5d585d06a8a req-ae20baf5-0cc5-4d3f-8be9-3f23dd796dca service nova] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Refreshing network info cache for port f4801f4b-2b95-41b6-a74c-7c95d83c7152 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 809.462370] env[63355]: DEBUG nova.network.neutron [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.709172] env[63355]: DEBUG nova.network.neutron [-] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.797548] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.798142] env[63355]: ERROR nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fcd0ef62-5937-42f8-ae04-94e0e375456e, please check neutron logs for more information. [ 809.798142] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Traceback (most recent call last): [ 809.798142] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 809.798142] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] self.driver.spawn(context, instance, image_meta, [ 809.798142] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 809.798142] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 809.798142] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 809.798142] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] vm_ref = self.build_virtual_machine(instance, [ 809.798142] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 809.798142] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] vif_infos = vmwarevif.get_vif_info(self._session, [ 809.798142] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] for vif in network_info: [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] return self._sync_wrapper(fn, *args, **kwargs) [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] self.wait() [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] self[:] = self._gt.wait() [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] return self._exit_event.wait() [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] result = hub.switch() [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 809.798543] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] return self.greenlet.switch() [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] result = function(*args, **kwargs) [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] return func(*args, **kwargs) [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] raise e [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] nwinfo = self.network_api.allocate_for_instance( [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] created_port_ids = self._update_ports_for_instance( [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] with excutils.save_and_reraise_exception(): [ 809.799063] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] self.force_reraise() [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] raise self.value [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] updated_port = self._update_port( [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] _ensure_no_port_binding_failure(port) [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] raise exception.PortBindingFailed(port_id=port['id']) [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] nova.exception.PortBindingFailed: Binding failed for port fcd0ef62-5937-42f8-ae04-94e0e375456e, please check neutron logs for more information. [ 809.799793] env[63355]: ERROR nova.compute.manager [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] [ 809.800244] env[63355]: DEBUG nova.compute.utils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Binding failed for port fcd0ef62-5937-42f8-ae04-94e0e375456e, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 809.800244] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.490s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.801423] env[63355]: INFO nova.compute.claims [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.803957] env[63355]: DEBUG nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Build of instance 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e was re-scheduled: Binding failed for port fcd0ef62-5937-42f8-ae04-94e0e375456e, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 809.804393] env[63355]: DEBUG nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 809.804611] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Acquiring lock "refresh_cache-05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.804769] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Acquired lock "refresh_cache-05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.804949] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.884714] env[63355]: DEBUG nova.network.neutron [req-6a856bd6-e970-408f-aa27-e5d585d06a8a req-ae20baf5-0cc5-4d3f-8be9-3f23dd796dca service nova] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.964815] env[63355]: INFO nova.compute.manager [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] [instance: 3c04c546-a9fa-4cbd-98d4-3ea92918a918] Took 1.02 seconds to deallocate network for instance. [ 809.974411] env[63355]: DEBUG nova.network.neutron [req-6a856bd6-e970-408f-aa27-e5d585d06a8a req-ae20baf5-0cc5-4d3f-8be9-3f23dd796dca service nova] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.212094] env[63355]: INFO nova.compute.manager [-] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Took 1.02 seconds to deallocate network for instance. [ 810.215121] env[63355]: DEBUG nova.compute.claims [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 810.215262] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.324161] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.391628] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.477440] env[63355]: DEBUG oslo_concurrency.lockutils [req-6a856bd6-e970-408f-aa27-e5d585d06a8a req-ae20baf5-0cc5-4d3f-8be9-3f23dd796dca service nova] Releasing lock "refresh_cache-7816d808-c9dd-403f-b993-c6d9dc458c7b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.477678] env[63355]: DEBUG nova.compute.manager [req-6a856bd6-e970-408f-aa27-e5d585d06a8a req-ae20baf5-0cc5-4d3f-8be9-3f23dd796dca service nova] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Received event network-vif-deleted-f4801f4b-2b95-41b6-a74c-7c95d83c7152 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.896197] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Releasing lock "refresh_cache-05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.896420] env[63355]: DEBUG nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 810.896600] env[63355]: DEBUG nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.896767] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 810.911692] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.990835] env[63355]: INFO nova.scheduler.client.report [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Deleted allocations for instance 3c04c546-a9fa-4cbd-98d4-3ea92918a918 [ 811.133304] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d8c6065-05f0-473e-8798-951c776ae1fd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.141686] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16759991-ab45-4ae2-9a10-55b57ef60143 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.172714] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9215dceb-3219-454c-9b87-477a13ec921a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.179683] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385fa843-39bc-4637-b40b-b5aab6e0db1e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.194699] env[63355]: DEBUG nova.compute.provider_tree [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.414175] env[63355]: DEBUG nova.network.neutron [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.499876] env[63355]: DEBUG oslo_concurrency.lockutils [None req-37c373ce-ed2f-4b04-b6f7-e8115e22a0a5 tempest-InstanceActionsTestJSON-362497317 tempest-InstanceActionsTestJSON-362497317-project-member] Lock "3c04c546-a9fa-4cbd-98d4-3ea92918a918" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 168.536s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.696865] env[63355]: DEBUG nova.scheduler.client.report [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.917042] env[63355]: INFO nova.compute.manager [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] [instance: 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e] Took 1.02 seconds to deallocate network for instance. [ 812.007314] env[63355]: DEBUG nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.201933] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.202501] env[63355]: DEBUG nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.205116] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.761s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.529694] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.713531] env[63355]: DEBUG nova.compute.utils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.718092] env[63355]: DEBUG nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.718450] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.789555] env[63355]: DEBUG nova.policy [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6eab8c9d2d08447b918d5b749b08fa86', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c31d477cfcf42f1b47c81438cf36855', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.954336] env[63355]: INFO nova.scheduler.client.report [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Deleted allocations for instance 05b9ea96-23f9-4c7a-aa0f-0d9142c4167e [ 813.118322] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baab1d4a-66b5-4d6f-86b8-c5122ae79954 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.125842] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df335b9b-6345-455f-a78e-276d24fe6389 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.156862] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Successfully created port: 3d179281-1687-4048-bd8a-8e574b2dd566 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.159564] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d2871f-c521-48ac-a2ab-2a3868a00e1e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.166681] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5582a549-707c-4844-b5af-4a315fe1c882 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.180412] env[63355]: DEBUG nova.compute.provider_tree [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.218620] env[63355]: DEBUG nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 813.465616] env[63355]: DEBUG oslo_concurrency.lockutils [None req-409c4358-2364-4260-9344-b78b7f2e91d3 tempest-ServersNegativeTestJSON-1737520908 tempest-ServersNegativeTestJSON-1737520908-project-member] Lock "05b9ea96-23f9-4c7a-aa0f-0d9142c4167e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 164.566s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.687551] env[63355]: DEBUG nova.scheduler.client.report [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.953824] env[63355]: DEBUG nova.compute.manager [req-72912319-4715-41c2-839b-5ece3ba6152e req-f01c4fe7-2ddc-4b1b-99a0-c00ecca56596 service nova] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Received event network-changed-3d179281-1687-4048-bd8a-8e574b2dd566 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.953959] env[63355]: DEBUG nova.compute.manager [req-72912319-4715-41c2-839b-5ece3ba6152e req-f01c4fe7-2ddc-4b1b-99a0-c00ecca56596 service nova] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Refreshing instance network info cache due to event network-changed-3d179281-1687-4048-bd8a-8e574b2dd566. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 813.954127] env[63355]: DEBUG oslo_concurrency.lockutils [req-72912319-4715-41c2-839b-5ece3ba6152e req-f01c4fe7-2ddc-4b1b-99a0-c00ecca56596 service nova] Acquiring lock "refresh_cache-a1e1e0ea-debd-4529-b175-4bdf44e4de31" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.954302] env[63355]: DEBUG oslo_concurrency.lockutils [req-72912319-4715-41c2-839b-5ece3ba6152e req-f01c4fe7-2ddc-4b1b-99a0-c00ecca56596 service nova] Acquired lock "refresh_cache-a1e1e0ea-debd-4529-b175-4bdf44e4de31" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.954462] env[63355]: DEBUG nova.network.neutron [req-72912319-4715-41c2-839b-5ece3ba6152e req-f01c4fe7-2ddc-4b1b-99a0-c00ecca56596 service nova] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Refreshing network info cache for port 3d179281-1687-4048-bd8a-8e574b2dd566 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 813.974023] env[63355]: DEBUG nova.compute.manager [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 814.199017] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.991s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.199017] env[63355]: ERROR nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 260b6560-883f-441f-ada4-45fef2d173d6, please check neutron logs for more information. [ 814.199017] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Traceback (most recent call last): [ 814.199017] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.199017] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] self.driver.spawn(context, instance, image_meta, [ 814.199017] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 814.199017] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.199017] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.199017] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] vm_ref = self.build_virtual_machine(instance, [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] for vif in network_info: [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] return self._sync_wrapper(fn, *args, **kwargs) [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] self.wait() [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] self[:] = self._gt.wait() [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] return self._exit_event.wait() [ 814.199405] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] current.throw(*self._exc) [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] result = function(*args, **kwargs) [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] return func(*args, **kwargs) [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] raise e [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] nwinfo = self.network_api.allocate_for_instance( [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] created_port_ids = self._update_ports_for_instance( [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.199786] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] with excutils.save_and_reraise_exception(): [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] self.force_reraise() [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] raise self.value [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] updated_port = self._update_port( [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] _ensure_no_port_binding_failure(port) [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] raise exception.PortBindingFailed(port_id=port['id']) [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] nova.exception.PortBindingFailed: Binding failed for port 260b6560-883f-441f-ada4-45fef2d173d6, please check neutron logs for more information. [ 814.200361] env[63355]: ERROR nova.compute.manager [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] [ 814.200760] env[63355]: DEBUG nova.compute.utils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Binding failed for port 260b6560-883f-441f-ada4-45fef2d173d6, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 814.204618] env[63355]: DEBUG nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Build of instance 282f1b55-8a74-467a-9ec8-0684eb780970 was re-scheduled: Binding failed for port 260b6560-883f-441f-ada4-45fef2d173d6, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 814.204618] env[63355]: DEBUG nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 814.204618] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Acquiring lock "refresh_cache-282f1b55-8a74-467a-9ec8-0684eb780970" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.204618] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Acquired lock "refresh_cache-282f1b55-8a74-467a-9ec8-0684eb780970" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.205102] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.205102] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.115s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.206046] env[63355]: INFO nova.compute.claims [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.227900] env[63355]: DEBUG nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 814.262511] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.263190] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.263513] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.264246] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.264903] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.265782] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.265782] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.265782] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.265925] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.266048] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.266202] env[63355]: DEBUG nova.virt.hardware [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.267251] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31798a95-4180-4abb-b93f-7d4a43670a07 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.276870] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8538dfba-cce9-45d9-a18a-24f399176c12 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.494881] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.503288] env[63355]: ERROR nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3d179281-1687-4048-bd8a-8e574b2dd566, please check neutron logs for more information. [ 814.503288] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 814.503288] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.503288] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 814.503288] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.503288] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 814.503288] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.503288] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 814.503288] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.503288] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 814.503288] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.503288] env[63355]: ERROR nova.compute.manager raise self.value [ 814.503288] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.503288] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 814.503288] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.503288] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 814.503893] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.503893] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 814.503893] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3d179281-1687-4048-bd8a-8e574b2dd566, please check neutron logs for more information. [ 814.503893] env[63355]: ERROR nova.compute.manager [ 814.503893] env[63355]: Traceback (most recent call last): [ 814.503893] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 814.503893] env[63355]: listener.cb(fileno) [ 814.503893] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.503893] env[63355]: result = function(*args, **kwargs) [ 814.503893] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 814.503893] env[63355]: return func(*args, **kwargs) [ 814.503893] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.503893] env[63355]: raise e [ 814.503893] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.503893] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 814.503893] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.503893] env[63355]: created_port_ids = self._update_ports_for_instance( [ 814.503893] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.503893] env[63355]: with excutils.save_and_reraise_exception(): [ 814.503893] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.503893] env[63355]: self.force_reraise() [ 814.503893] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.503893] env[63355]: raise self.value [ 814.503893] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.503893] env[63355]: updated_port = self._update_port( [ 814.503893] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.503893] env[63355]: _ensure_no_port_binding_failure(port) [ 814.503893] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.503893] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 814.505031] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 3d179281-1687-4048-bd8a-8e574b2dd566, please check neutron logs for more information. [ 814.505031] env[63355]: Removing descriptor: 17 [ 814.505031] env[63355]: ERROR nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3d179281-1687-4048-bd8a-8e574b2dd566, please check neutron logs for more information. [ 814.505031] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Traceback (most recent call last): [ 814.505031] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 814.505031] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] yield resources [ 814.505031] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.505031] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] self.driver.spawn(context, instance, image_meta, [ 814.505031] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 814.505031] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.505031] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.505031] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] vm_ref = self.build_virtual_machine(instance, [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] for vif in network_info: [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] return self._sync_wrapper(fn, *args, **kwargs) [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] self.wait() [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] self[:] = self._gt.wait() [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] return self._exit_event.wait() [ 814.505452] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] result = hub.switch() [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] return self.greenlet.switch() [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] result = function(*args, **kwargs) [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] return func(*args, **kwargs) [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] raise e [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] nwinfo = self.network_api.allocate_for_instance( [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.505932] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] created_port_ids = self._update_ports_for_instance( [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] with excutils.save_and_reraise_exception(): [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] self.force_reraise() [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] raise self.value [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] updated_port = self._update_port( [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] _ensure_no_port_binding_failure(port) [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.506398] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] raise exception.PortBindingFailed(port_id=port['id']) [ 814.506819] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] nova.exception.PortBindingFailed: Binding failed for port 3d179281-1687-4048-bd8a-8e574b2dd566, please check neutron logs for more information. [ 814.506819] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] [ 814.506819] env[63355]: INFO nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Terminating instance [ 814.507726] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquiring lock "refresh_cache-a1e1e0ea-debd-4529-b175-4bdf44e4de31" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.585337] env[63355]: DEBUG nova.network.neutron [req-72912319-4715-41c2-839b-5ece3ba6152e req-f01c4fe7-2ddc-4b1b-99a0-c00ecca56596 service nova] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.729537] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.781997] env[63355]: DEBUG nova.network.neutron [req-72912319-4715-41c2-839b-5ece3ba6152e req-f01c4fe7-2ddc-4b1b-99a0-c00ecca56596 service nova] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.842290] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.288269] env[63355]: DEBUG oslo_concurrency.lockutils [req-72912319-4715-41c2-839b-5ece3ba6152e req-f01c4fe7-2ddc-4b1b-99a0-c00ecca56596 service nova] Releasing lock "refresh_cache-a1e1e0ea-debd-4529-b175-4bdf44e4de31" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.288728] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquired lock "refresh_cache-a1e1e0ea-debd-4529-b175-4bdf44e4de31" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.288891] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.345304] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Releasing lock "refresh_cache-282f1b55-8a74-467a-9ec8-0684eb780970" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.345541] env[63355]: DEBUG nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 815.345728] env[63355]: DEBUG nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.345896] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.367328] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.680614] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e999ac5-fbc3-4e0a-ab6c-70e457e6aab6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.693152] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5535c8-1d2b-4a08-8039-462b2dfd0a09 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.725195] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc6dc99-e351-4f22-bfab-aec5e4abd71c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.732891] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d12a64-6764-473e-a6ea-3955d5d45eec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.750287] env[63355]: DEBUG nova.compute.provider_tree [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.812455] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.875114] env[63355]: DEBUG nova.network.neutron [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.921157] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.047268] env[63355]: DEBUG nova.compute.manager [req-9333ca83-de35-4fbe-802c-ab06662d945a req-a3e8b8f8-334f-4bb5-a559-8f0724fe26e5 service nova] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Received event network-vif-deleted-3d179281-1687-4048-bd8a-8e574b2dd566 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.254197] env[63355]: DEBUG nova.scheduler.client.report [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.378297] env[63355]: INFO nova.compute.manager [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] [instance: 282f1b55-8a74-467a-9ec8-0684eb780970] Took 1.03 seconds to deallocate network for instance. [ 816.425339] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Releasing lock "refresh_cache-a1e1e0ea-debd-4529-b175-4bdf44e4de31" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.425705] env[63355]: DEBUG nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 816.426074] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 816.426705] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7137ea35-5733-4c97-b162-c6b55a5e4931 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.436490] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66271838-30bc-45ec-b090-736f20e9a562 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.458874] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a1e1e0ea-debd-4529-b175-4bdf44e4de31 could not be found. [ 816.459153] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.459341] env[63355]: INFO nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Took 0.03 seconds to destroy the instance on the hypervisor. [ 816.459584] env[63355]: DEBUG oslo.service.loopingcall [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.459793] env[63355]: DEBUG nova.compute.manager [-] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 816.459886] env[63355]: DEBUG nova.network.neutron [-] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.477266] env[63355]: DEBUG nova.network.neutron [-] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.759792] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.760344] env[63355]: DEBUG nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.763130] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.731s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.764671] env[63355]: INFO nova.compute.claims [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.980144] env[63355]: DEBUG nova.network.neutron [-] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.270558] env[63355]: DEBUG nova.compute.utils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.275569] env[63355]: DEBUG nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.275569] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.326852] env[63355]: DEBUG nova.policy [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfbc8f0af60940eeb0617698fb003a02', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd36c7cc647f741b38d717ba7c6458088', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.413582] env[63355]: INFO nova.scheduler.client.report [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Deleted allocations for instance 282f1b55-8a74-467a-9ec8-0684eb780970 [ 817.483266] env[63355]: INFO nova.compute.manager [-] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Took 1.02 seconds to deallocate network for instance. [ 817.485761] env[63355]: DEBUG nova.compute.claims [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 817.485942] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.686061] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Successfully created port: 7113633c-13bf-4d05-a697-7df302bda34e {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.777321] env[63355]: DEBUG nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.921694] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0bccd3f2-0ce5-4054-8db6-4c328fbf8f2e tempest-ServersTestBootFromVolume-1461075465 tempest-ServersTestBootFromVolume-1461075465-project-member] Lock "282f1b55-8a74-467a-9ec8-0684eb780970" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 168.419s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.263032] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0685b755-af85-4d1a-a724-9fef0e2bf666 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.272156] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d98e28-c20c-4a44-b99a-c18f9bb625bf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.305785] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a08a33-7a67-4c5a-b548-f2858955b054 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.314689] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83565a6-674d-430f-8d44-278332606ef7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.337030] env[63355]: DEBUG nova.compute.provider_tree [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.424724] env[63355]: DEBUG nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.472608] env[63355]: DEBUG nova.compute.manager [req-a0d0a837-d035-4d99-bec1-fe0383b855d6 req-5d16023c-ca9c-4093-8708-c36d56952d67 service nova] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Received event network-changed-7113633c-13bf-4d05-a697-7df302bda34e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.476020] env[63355]: DEBUG nova.compute.manager [req-a0d0a837-d035-4d99-bec1-fe0383b855d6 req-5d16023c-ca9c-4093-8708-c36d56952d67 service nova] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Refreshing instance network info cache due to event network-changed-7113633c-13bf-4d05-a697-7df302bda34e. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 818.476402] env[63355]: DEBUG oslo_concurrency.lockutils [req-a0d0a837-d035-4d99-bec1-fe0383b855d6 req-5d16023c-ca9c-4093-8708-c36d56952d67 service nova] Acquiring lock "refresh_cache-21e17d32-40db-4dc1-9e73-fa6d35a4c06a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.476724] env[63355]: DEBUG oslo_concurrency.lockutils [req-a0d0a837-d035-4d99-bec1-fe0383b855d6 req-5d16023c-ca9c-4093-8708-c36d56952d67 service nova] Acquired lock "refresh_cache-21e17d32-40db-4dc1-9e73-fa6d35a4c06a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.476985] env[63355]: DEBUG nova.network.neutron [req-a0d0a837-d035-4d99-bec1-fe0383b855d6 req-5d16023c-ca9c-4093-8708-c36d56952d67 service nova] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Refreshing network info cache for port 7113633c-13bf-4d05-a697-7df302bda34e {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.689569] env[63355]: ERROR nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7113633c-13bf-4d05-a697-7df302bda34e, please check neutron logs for more information. [ 818.689569] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 818.689569] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.689569] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 818.689569] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 818.689569] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 818.689569] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 818.689569] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 818.689569] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.689569] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 818.689569] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.689569] env[63355]: ERROR nova.compute.manager raise self.value [ 818.689569] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 818.689569] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 818.689569] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.689569] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 818.690195] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.690195] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 818.690195] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7113633c-13bf-4d05-a697-7df302bda34e, please check neutron logs for more information. [ 818.690195] env[63355]: ERROR nova.compute.manager [ 818.690195] env[63355]: Traceback (most recent call last): [ 818.690195] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 818.690195] env[63355]: listener.cb(fileno) [ 818.690195] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 818.690195] env[63355]: result = function(*args, **kwargs) [ 818.690195] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 818.690195] env[63355]: return func(*args, **kwargs) [ 818.690195] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 818.690195] env[63355]: raise e [ 818.690195] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.690195] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 818.690195] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 818.690195] env[63355]: created_port_ids = self._update_ports_for_instance( [ 818.690195] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 818.690195] env[63355]: with excutils.save_and_reraise_exception(): [ 818.690195] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.690195] env[63355]: self.force_reraise() [ 818.690195] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.690195] env[63355]: raise self.value [ 818.690195] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 818.690195] env[63355]: updated_port = self._update_port( [ 818.690195] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.690195] env[63355]: _ensure_no_port_binding_failure(port) [ 818.690195] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.690195] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 818.691249] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 7113633c-13bf-4d05-a697-7df302bda34e, please check neutron logs for more information. [ 818.691249] env[63355]: Removing descriptor: 17 [ 818.810730] env[63355]: DEBUG nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.838341] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.838584] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.838737] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.838910] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.839064] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.839210] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.839411] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.839564] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.839726] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.839883] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.841211] env[63355]: DEBUG nova.virt.hardware [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.841947] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13297df-58c7-434a-ba11-075ad4c6e17c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.845276] env[63355]: DEBUG nova.scheduler.client.report [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.855700] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e38cee-02b5-4eaf-872b-b9fdde43e45b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.870326] env[63355]: ERROR nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7113633c-13bf-4d05-a697-7df302bda34e, please check neutron logs for more information. [ 818.870326] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Traceback (most recent call last): [ 818.870326] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 818.870326] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] yield resources [ 818.870326] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 818.870326] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] self.driver.spawn(context, instance, image_meta, [ 818.870326] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 818.870326] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 818.870326] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 818.870326] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] vm_ref = self.build_virtual_machine(instance, [ 818.870326] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] vif_infos = vmwarevif.get_vif_info(self._session, [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] for vif in network_info: [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] return self._sync_wrapper(fn, *args, **kwargs) [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] self.wait() [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] self[:] = self._gt.wait() [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] return self._exit_event.wait() [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 818.870707] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] current.throw(*self._exc) [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] result = function(*args, **kwargs) [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] return func(*args, **kwargs) [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] raise e [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] nwinfo = self.network_api.allocate_for_instance( [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] created_port_ids = self._update_ports_for_instance( [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] with excutils.save_and_reraise_exception(): [ 818.871236] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] self.force_reraise() [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] raise self.value [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] updated_port = self._update_port( [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] _ensure_no_port_binding_failure(port) [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] raise exception.PortBindingFailed(port_id=port['id']) [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] nova.exception.PortBindingFailed: Binding failed for port 7113633c-13bf-4d05-a697-7df302bda34e, please check neutron logs for more information. [ 818.875424] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] [ 818.875424] env[63355]: INFO nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Terminating instance [ 818.876953] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "refresh_cache-21e17d32-40db-4dc1-9e73-fa6d35a4c06a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.964606] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.997338] env[63355]: DEBUG nova.network.neutron [req-a0d0a837-d035-4d99-bec1-fe0383b855d6 req-5d16023c-ca9c-4093-8708-c36d56952d67 service nova] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.088787] env[63355]: DEBUG nova.network.neutron [req-a0d0a837-d035-4d99-bec1-fe0383b855d6 req-5d16023c-ca9c-4093-8708-c36d56952d67 service nova] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.350903] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.351495] env[63355]: DEBUG nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.356755] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.638s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.592078] env[63355]: DEBUG oslo_concurrency.lockutils [req-a0d0a837-d035-4d99-bec1-fe0383b855d6 req-5d16023c-ca9c-4093-8708-c36d56952d67 service nova] Releasing lock "refresh_cache-21e17d32-40db-4dc1-9e73-fa6d35a4c06a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.592078] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquired lock "refresh_cache-21e17d32-40db-4dc1-9e73-fa6d35a4c06a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.592078] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.865506] env[63355]: DEBUG nova.compute.utils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.870436] env[63355]: DEBUG nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 819.870679] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 819.939881] env[63355]: DEBUG nova.policy [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6eab8c9d2d08447b918d5b749b08fa86', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c31d477cfcf42f1b47c81438cf36855', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.113526] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.247533] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.270018] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035d025b-bb6b-4cda-997b-1a2dc6d8b0ca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.277059] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03808feb-836d-44cb-b702-53a17becdbc0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.310998] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ae3062-d525-4343-af69-c7e6ea2cd93f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.318470] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bde2d0a-19c0-49af-8e21-f2fbcbbe747f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.331290] env[63355]: DEBUG nova.compute.provider_tree [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.374504] env[63355]: DEBUG nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.409841] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Successfully created port: 7d88a38c-9dc5-4b16-bcee-211885563c6a {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.504592] env[63355]: DEBUG nova.compute.manager [req-f49e0a9f-1424-41ec-89ea-57264166cca9 req-5b0326a3-757a-4ca7-a276-b66767a07d02 service nova] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Received event network-vif-deleted-7113633c-13bf-4d05-a697-7df302bda34e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.752245] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Releasing lock "refresh_cache-21e17d32-40db-4dc1-9e73-fa6d35a4c06a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.753720] env[63355]: DEBUG nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 820.753720] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.753720] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71b21789-54d5-4208-aed1-7f5c94fe8721 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.767289] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fa3ac8-b148-4e8f-aa36-8b5c7d674def {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.790574] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 21e17d32-40db-4dc1-9e73-fa6d35a4c06a could not be found. [ 820.790839] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.790993] env[63355]: INFO nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 820.791356] env[63355]: DEBUG oslo.service.loopingcall [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.791594] env[63355]: DEBUG nova.compute.manager [-] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.791839] env[63355]: DEBUG nova.network.neutron [-] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 820.829436] env[63355]: DEBUG nova.network.neutron [-] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.834518] env[63355]: DEBUG nova.scheduler.client.report [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.335373] env[63355]: DEBUG nova.network.neutron [-] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.341820] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.985s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.342572] env[63355]: ERROR nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8b6ff543-f0de-4659-8c3f-93517f56fc21, please check neutron logs for more information. [ 821.342572] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Traceback (most recent call last): [ 821.342572] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 821.342572] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] self.driver.spawn(context, instance, image_meta, [ 821.342572] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 821.342572] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 821.342572] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 821.342572] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] vm_ref = self.build_virtual_machine(instance, [ 821.342572] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 821.342572] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 821.342572] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] for vif in network_info: [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] return self._sync_wrapper(fn, *args, **kwargs) [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] self.wait() [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] self[:] = self._gt.wait() [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] return self._exit_event.wait() [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] current.throw(*self._exc) [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.343088] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] result = function(*args, **kwargs) [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] return func(*args, **kwargs) [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] raise e [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] nwinfo = self.network_api.allocate_for_instance( [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] created_port_ids = self._update_ports_for_instance( [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] with excutils.save_and_reraise_exception(): [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] self.force_reraise() [ 821.345076] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.345353] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] raise self.value [ 821.345353] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.345353] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] updated_port = self._update_port( [ 821.345353] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.345353] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] _ensure_no_port_binding_failure(port) [ 821.345353] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.345353] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] raise exception.PortBindingFailed(port_id=port['id']) [ 821.345353] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] nova.exception.PortBindingFailed: Binding failed for port 8b6ff543-f0de-4659-8c3f-93517f56fc21, please check neutron logs for more information. [ 821.345353] env[63355]: ERROR nova.compute.manager [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] [ 821.345353] env[63355]: DEBUG nova.compute.utils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Binding failed for port 8b6ff543-f0de-4659-8c3f-93517f56fc21, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 821.345549] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.796s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.347143] env[63355]: INFO nova.compute.claims [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.351375] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Build of instance 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d was re-scheduled: Binding failed for port 8b6ff543-f0de-4659-8c3f-93517f56fc21, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 821.351823] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 821.352072] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "refresh_cache-7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.352229] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired lock "refresh_cache-7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.352395] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.388025] env[63355]: DEBUG nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.433841] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.435009] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.435009] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.435009] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.435009] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.435009] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.435170] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.435197] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.435363] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.435521] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.435684] env[63355]: DEBUG nova.virt.hardware [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.436884] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee4c008-93c9-4a38-91b9-aa6475830e1c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.445151] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e831246b-70c4-47c6-aab7-adb3d6dc9f86 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.696929] env[63355]: ERROR nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7d88a38c-9dc5-4b16-bcee-211885563c6a, please check neutron logs for more information. [ 821.696929] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 821.696929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.696929] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 821.696929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.696929] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 821.696929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.696929] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 821.696929] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.696929] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 821.696929] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.696929] env[63355]: ERROR nova.compute.manager raise self.value [ 821.696929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.696929] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 821.696929] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.696929] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 821.697355] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.697355] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 821.697355] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7d88a38c-9dc5-4b16-bcee-211885563c6a, please check neutron logs for more information. [ 821.697355] env[63355]: ERROR nova.compute.manager [ 821.697355] env[63355]: Traceback (most recent call last): [ 821.697355] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 821.697355] env[63355]: listener.cb(fileno) [ 821.697355] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.697355] env[63355]: result = function(*args, **kwargs) [ 821.697355] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 821.697355] env[63355]: return func(*args, **kwargs) [ 821.697355] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.697355] env[63355]: raise e [ 821.697355] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.697355] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 821.697355] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.697355] env[63355]: created_port_ids = self._update_ports_for_instance( [ 821.697355] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.697355] env[63355]: with excutils.save_and_reraise_exception(): [ 821.697355] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.697355] env[63355]: self.force_reraise() [ 821.697355] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.697355] env[63355]: raise self.value [ 821.697355] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.697355] env[63355]: updated_port = self._update_port( [ 821.697355] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.697355] env[63355]: _ensure_no_port_binding_failure(port) [ 821.697355] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.697355] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 821.698153] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 7d88a38c-9dc5-4b16-bcee-211885563c6a, please check neutron logs for more information. [ 821.698153] env[63355]: Removing descriptor: 17 [ 821.698153] env[63355]: ERROR nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7d88a38c-9dc5-4b16-bcee-211885563c6a, please check neutron logs for more information. [ 821.698153] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Traceback (most recent call last): [ 821.698153] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 821.698153] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] yield resources [ 821.698153] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 821.698153] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] self.driver.spawn(context, instance, image_meta, [ 821.698153] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 821.698153] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] self._vmops.spawn(context, instance, image_meta, injected_files, [ 821.698153] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 821.698153] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] vm_ref = self.build_virtual_machine(instance, [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] vif_infos = vmwarevif.get_vif_info(self._session, [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] for vif in network_info: [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] return self._sync_wrapper(fn, *args, **kwargs) [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] self.wait() [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] self[:] = self._gt.wait() [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] return self._exit_event.wait() [ 821.698454] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] result = hub.switch() [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] return self.greenlet.switch() [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] result = function(*args, **kwargs) [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] return func(*args, **kwargs) [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] raise e [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] nwinfo = self.network_api.allocate_for_instance( [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.698806] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] created_port_ids = self._update_ports_for_instance( [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] with excutils.save_and_reraise_exception(): [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] self.force_reraise() [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] raise self.value [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] updated_port = self._update_port( [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] _ensure_no_port_binding_failure(port) [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.699129] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] raise exception.PortBindingFailed(port_id=port['id']) [ 821.699409] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] nova.exception.PortBindingFailed: Binding failed for port 7d88a38c-9dc5-4b16-bcee-211885563c6a, please check neutron logs for more information. [ 821.699409] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] [ 821.699409] env[63355]: INFO nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Terminating instance [ 821.700275] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquiring lock "refresh_cache-7d79c9cf-66a9-4117-b090-0ab0676c3114" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.700434] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquired lock "refresh_cache-7d79c9cf-66a9-4117-b090-0ab0676c3114" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.700598] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.838597] env[63355]: INFO nova.compute.manager [-] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Took 1.05 seconds to deallocate network for instance. [ 821.841285] env[63355]: DEBUG nova.compute.claims [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 821.841466] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.874475] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.953693] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.220012] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.297917] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.456962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Releasing lock "refresh_cache-7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.457229] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 822.457426] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.457634] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 822.478422] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.542215] env[63355]: DEBUG nova.compute.manager [req-72d415d2-27ed-49bb-b812-76acc35eef80 req-546c886d-e19c-4ad6-a739-adb72746f386 service nova] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Received event network-changed-7d88a38c-9dc5-4b16-bcee-211885563c6a {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.542378] env[63355]: DEBUG nova.compute.manager [req-72d415d2-27ed-49bb-b812-76acc35eef80 req-546c886d-e19c-4ad6-a739-adb72746f386 service nova] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Refreshing instance network info cache due to event network-changed-7d88a38c-9dc5-4b16-bcee-211885563c6a. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.542558] env[63355]: DEBUG oslo_concurrency.lockutils [req-72d415d2-27ed-49bb-b812-76acc35eef80 req-546c886d-e19c-4ad6-a739-adb72746f386 service nova] Acquiring lock "refresh_cache-7d79c9cf-66a9-4117-b090-0ab0676c3114" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.700266] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34a7d26-1404-441d-b4ff-8e668b3541f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.708671] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd39f29-c452-44fd-8671-912157a9cf4a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.740878] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72da6e2-fad1-42f9-8b6b-d3d6d9ca46a2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.746612] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b76bcda-ab2c-4447-b7ab-1bc8df648d2d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.761133] env[63355]: DEBUG nova.compute.provider_tree [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.801536] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Releasing lock "refresh_cache-7d79c9cf-66a9-4117-b090-0ab0676c3114" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.802225] env[63355]: DEBUG nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 822.802225] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 822.802766] env[63355]: DEBUG oslo_concurrency.lockutils [req-72d415d2-27ed-49bb-b812-76acc35eef80 req-546c886d-e19c-4ad6-a739-adb72746f386 service nova] Acquired lock "refresh_cache-7d79c9cf-66a9-4117-b090-0ab0676c3114" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.803104] env[63355]: DEBUG nova.network.neutron [req-72d415d2-27ed-49bb-b812-76acc35eef80 req-546c886d-e19c-4ad6-a739-adb72746f386 service nova] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Refreshing network info cache for port 7d88a38c-9dc5-4b16-bcee-211885563c6a {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.804185] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d2e8cbb-8bec-4c42-bc15-ad0bedb89938 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.814358] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20a60f0-3842-46ec-b2af-c040b1c0799b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.835670] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7d79c9cf-66a9-4117-b090-0ab0676c3114 could not be found. [ 822.835897] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 822.836174] env[63355]: INFO nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Took 0.03 seconds to destroy the instance on the hypervisor. [ 822.836417] env[63355]: DEBUG oslo.service.loopingcall [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.836618] env[63355]: DEBUG nova.compute.manager [-] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.836708] env[63355]: DEBUG nova.network.neutron [-] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 822.851848] env[63355]: DEBUG nova.network.neutron [-] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.981453] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.264447] env[63355]: DEBUG nova.scheduler.client.report [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.322820] env[63355]: DEBUG nova.network.neutron [req-72d415d2-27ed-49bb-b812-76acc35eef80 req-546c886d-e19c-4ad6-a739-adb72746f386 service nova] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.354522] env[63355]: DEBUG nova.network.neutron [-] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.392882] env[63355]: DEBUG nova.network.neutron [req-72d415d2-27ed-49bb-b812-76acc35eef80 req-546c886d-e19c-4ad6-a739-adb72746f386 service nova] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.484987] env[63355]: INFO nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d] Took 1.03 seconds to deallocate network for instance. [ 823.769350] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.769867] env[63355]: DEBUG nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.772428] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.033s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.772602] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.772759] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 823.773094] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.008s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.776300] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00fdc55d-26e6-4d84-a4ba-96ed5b6bce09 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.785957] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c061396d-2cbc-4865-b090-8fb1d3d18c66 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.799274] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee594a9-4eb9-47cb-b907-769d9dbb8323 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.806277] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401112a7-b805-4338-8e8b-c0bde7dfb50a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.835908] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181519MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 823.836055] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.860976] env[63355]: INFO nova.compute.manager [-] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Took 1.02 seconds to deallocate network for instance. [ 823.863361] env[63355]: DEBUG nova.compute.claims [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 823.863539] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.895602] env[63355]: DEBUG oslo_concurrency.lockutils [req-72d415d2-27ed-49bb-b812-76acc35eef80 req-546c886d-e19c-4ad6-a739-adb72746f386 service nova] Releasing lock "refresh_cache-7d79c9cf-66a9-4117-b090-0ab0676c3114" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.895870] env[63355]: DEBUG nova.compute.manager [req-72d415d2-27ed-49bb-b812-76acc35eef80 req-546c886d-e19c-4ad6-a739-adb72746f386 service nova] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Received event network-vif-deleted-7d88a38c-9dc5-4b16-bcee-211885563c6a {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.277205] env[63355]: DEBUG nova.compute.utils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.279144] env[63355]: DEBUG nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.279325] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 824.320693] env[63355]: DEBUG nova.policy [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd32aea4219944964ba3fa6eed5d70ea7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58271233c0bc4384acb7cd581d3bed59', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.522799] env[63355]: INFO nova.scheduler.client.report [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Deleted allocations for instance 7d7519ce-db6a-416b-b7ee-b5aa3364fa4d [ 824.605814] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Successfully created port: 098f7509-6158-4f8f-bc97-63435fa49c6f {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 824.683766] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d8f986b-f7a9-403e-9c84-1af107c96764 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.693143] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0bdf418-2129-4041-9163-6572cae6c81f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.723265] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1582d144-90a5-4127-8e4d-3f4907db2a6f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.730742] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db030738-7171-43c6-a410-d9afa804ba94 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.745338] env[63355]: DEBUG nova.compute.provider_tree [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.782472] env[63355]: DEBUG nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 825.033674] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "7d7519ce-db6a-416b-b7ee-b5aa3364fa4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 170.490s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.247988] env[63355]: DEBUG nova.scheduler.client.report [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.256083] env[63355]: DEBUG nova.compute.manager [req-c54bbb61-9a18-4f28-b66f-8bc1234f1af6 req-f212e12e-b0d1-44b1-9627-70d86f6809e3 service nova] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Received event network-changed-098f7509-6158-4f8f-bc97-63435fa49c6f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.256285] env[63355]: DEBUG nova.compute.manager [req-c54bbb61-9a18-4f28-b66f-8bc1234f1af6 req-f212e12e-b0d1-44b1-9627-70d86f6809e3 service nova] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Refreshing instance network info cache due to event network-changed-098f7509-6158-4f8f-bc97-63435fa49c6f. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 825.256489] env[63355]: DEBUG oslo_concurrency.lockutils [req-c54bbb61-9a18-4f28-b66f-8bc1234f1af6 req-f212e12e-b0d1-44b1-9627-70d86f6809e3 service nova] Acquiring lock "refresh_cache-4ded6053-9afc-444e-bdfc-244265ed19a1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.256628] env[63355]: DEBUG oslo_concurrency.lockutils [req-c54bbb61-9a18-4f28-b66f-8bc1234f1af6 req-f212e12e-b0d1-44b1-9627-70d86f6809e3 service nova] Acquired lock "refresh_cache-4ded6053-9afc-444e-bdfc-244265ed19a1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.257073] env[63355]: DEBUG nova.network.neutron [req-c54bbb61-9a18-4f28-b66f-8bc1234f1af6 req-f212e12e-b0d1-44b1-9627-70d86f6809e3 service nova] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Refreshing network info cache for port 098f7509-6158-4f8f-bc97-63435fa49c6f {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.441208] env[63355]: ERROR nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 098f7509-6158-4f8f-bc97-63435fa49c6f, please check neutron logs for more information. [ 825.441208] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 825.441208] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 825.441208] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 825.441208] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 825.441208] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 825.441208] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 825.441208] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 825.441208] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 825.441208] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 825.441208] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 825.441208] env[63355]: ERROR nova.compute.manager raise self.value [ 825.441208] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 825.441208] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 825.441208] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 825.441208] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 825.441527] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 825.441527] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 825.441527] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 098f7509-6158-4f8f-bc97-63435fa49c6f, please check neutron logs for more information. [ 825.441527] env[63355]: ERROR nova.compute.manager [ 825.441527] env[63355]: Traceback (most recent call last): [ 825.441527] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 825.441527] env[63355]: listener.cb(fileno) [ 825.441527] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 825.441527] env[63355]: result = function(*args, **kwargs) [ 825.441527] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 825.441527] env[63355]: return func(*args, **kwargs) [ 825.441527] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 825.441527] env[63355]: raise e [ 825.441527] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 825.441527] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 825.441527] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 825.441527] env[63355]: created_port_ids = self._update_ports_for_instance( [ 825.441527] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 825.441527] env[63355]: with excutils.save_and_reraise_exception(): [ 825.441527] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 825.441527] env[63355]: self.force_reraise() [ 825.441527] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 825.441527] env[63355]: raise self.value [ 825.441527] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 825.441527] env[63355]: updated_port = self._update_port( [ 825.441527] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 825.441527] env[63355]: _ensure_no_port_binding_failure(port) [ 825.441527] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 825.441527] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 825.442106] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 098f7509-6158-4f8f-bc97-63435fa49c6f, please check neutron logs for more information. [ 825.442106] env[63355]: Removing descriptor: 17 [ 825.535924] env[63355]: DEBUG nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.760162] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.987s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.760669] env[63355]: ERROR nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port af5b9176-6396-4c83-8612-ed8c639cbcd4, please check neutron logs for more information. [ 825.760669] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Traceback (most recent call last): [ 825.760669] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 825.760669] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] self.driver.spawn(context, instance, image_meta, [ 825.760669] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 825.760669] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] self._vmops.spawn(context, instance, image_meta, injected_files, [ 825.760669] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 825.760669] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] vm_ref = self.build_virtual_machine(instance, [ 825.760669] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 825.760669] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] vif_infos = vmwarevif.get_vif_info(self._session, [ 825.760669] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] for vif in network_info: [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] return self._sync_wrapper(fn, *args, **kwargs) [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] self.wait() [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] self[:] = self._gt.wait() [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] return self._exit_event.wait() [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] result = hub.switch() [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 825.760956] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] return self.greenlet.switch() [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] result = function(*args, **kwargs) [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] return func(*args, **kwargs) [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] raise e [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] nwinfo = self.network_api.allocate_for_instance( [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] created_port_ids = self._update_ports_for_instance( [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] with excutils.save_and_reraise_exception(): [ 825.761290] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] self.force_reraise() [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] raise self.value [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] updated_port = self._update_port( [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] _ensure_no_port_binding_failure(port) [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] raise exception.PortBindingFailed(port_id=port['id']) [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] nova.exception.PortBindingFailed: Binding failed for port af5b9176-6396-4c83-8612-ed8c639cbcd4, please check neutron logs for more information. [ 825.761618] env[63355]: ERROR nova.compute.manager [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] [ 825.761992] env[63355]: DEBUG nova.compute.utils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Binding failed for port af5b9176-6396-4c83-8612-ed8c639cbcd4, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 825.765139] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Build of instance 795b61f6-a0f3-4aab-94b3-2540ccbb7312 was re-scheduled: Binding failed for port af5b9176-6396-4c83-8612-ed8c639cbcd4, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 825.765633] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 825.765921] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "refresh_cache-795b61f6-a0f3-4aab-94b3-2540ccbb7312" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.766105] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired lock "refresh_cache-795b61f6-a0f3-4aab-94b3-2540ccbb7312" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.766310] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.767349] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.552s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.785970] env[63355]: DEBUG nova.network.neutron [req-c54bbb61-9a18-4f28-b66f-8bc1234f1af6 req-f212e12e-b0d1-44b1-9627-70d86f6809e3 service nova] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.791830] env[63355]: DEBUG nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.817628] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.817873] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.818037] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.818225] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.818369] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.818513] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.818713] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.818868] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.819041] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.819206] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.819375] env[63355]: DEBUG nova.virt.hardware [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.820467] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c64dc3-2dde-470f-a587-cfa2252a3bec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.829959] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0558264-2de8-43f5-ac6a-cef925e16ad3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.843461] env[63355]: ERROR nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 098f7509-6158-4f8f-bc97-63435fa49c6f, please check neutron logs for more information. [ 825.843461] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Traceback (most recent call last): [ 825.843461] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 825.843461] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] yield resources [ 825.843461] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 825.843461] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] self.driver.spawn(context, instance, image_meta, [ 825.843461] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 825.843461] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 825.843461] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 825.843461] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] vm_ref = self.build_virtual_machine(instance, [ 825.843461] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] vif_infos = vmwarevif.get_vif_info(self._session, [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] for vif in network_info: [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] return self._sync_wrapper(fn, *args, **kwargs) [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] self.wait() [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] self[:] = self._gt.wait() [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] return self._exit_event.wait() [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 825.843776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] current.throw(*self._exc) [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] result = function(*args, **kwargs) [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] return func(*args, **kwargs) [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] raise e [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] nwinfo = self.network_api.allocate_for_instance( [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] created_port_ids = self._update_ports_for_instance( [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] with excutils.save_and_reraise_exception(): [ 825.844106] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] self.force_reraise() [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] raise self.value [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] updated_port = self._update_port( [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] _ensure_no_port_binding_failure(port) [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] raise exception.PortBindingFailed(port_id=port['id']) [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] nova.exception.PortBindingFailed: Binding failed for port 098f7509-6158-4f8f-bc97-63435fa49c6f, please check neutron logs for more information. [ 825.844469] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] [ 825.844469] env[63355]: INFO nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Terminating instance [ 825.847789] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquiring lock "refresh_cache-4ded6053-9afc-444e-bdfc-244265ed19a1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.886912] env[63355]: DEBUG nova.network.neutron [req-c54bbb61-9a18-4f28-b66f-8bc1234f1af6 req-f212e12e-b0d1-44b1-9627-70d86f6809e3 service nova] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.058073] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.290815] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.369661] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.389956] env[63355]: DEBUG oslo_concurrency.lockutils [req-c54bbb61-9a18-4f28-b66f-8bc1234f1af6 req-f212e12e-b0d1-44b1-9627-70d86f6809e3 service nova] Releasing lock "refresh_cache-4ded6053-9afc-444e-bdfc-244265ed19a1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.390567] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquired lock "refresh_cache-4ded6053-9afc-444e-bdfc-244265ed19a1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.391317] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.545243] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2dd2d2-9af7-4a16-b286-966237fed7ec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.552983] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e049f2-423f-444c-8f31-ee723f0e582a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.582683] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea14acb1-c7c9-4fe2-84b4-2c24e497ada5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.589825] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebd2fd5-9521-4b6e-83e6-24a649a7028c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.603557] env[63355]: DEBUG nova.compute.provider_tree [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.874133] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Releasing lock "refresh_cache-795b61f6-a0f3-4aab-94b3-2540ccbb7312" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.874133] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 826.874133] env[63355]: DEBUG nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.874133] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 826.888618] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.905236] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.975062] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.106965] env[63355]: DEBUG nova.scheduler.client.report [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.279079] env[63355]: DEBUG nova.compute.manager [req-f8c30676-f29e-4a4f-b3b3-b9511ec010eb req-7e4f804a-d886-464e-b4d2-8aba803c164b service nova] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Received event network-vif-deleted-098f7509-6158-4f8f-bc97-63435fa49c6f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.391520] env[63355]: DEBUG nova.network.neutron [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.477607] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Releasing lock "refresh_cache-4ded6053-9afc-444e-bdfc-244265ed19a1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.478085] env[63355]: DEBUG nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.478337] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 827.478641] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6268f776-334d-42fe-861d-7c6278cef1b3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.488266] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd4257e-d4de-4030-960a-eabc95504a8c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.508651] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4ded6053-9afc-444e-bdfc-244265ed19a1 could not be found. [ 827.508860] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 827.509053] env[63355]: INFO nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 827.509299] env[63355]: DEBUG oslo.service.loopingcall [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.509504] env[63355]: DEBUG nova.compute.manager [-] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 827.509594] env[63355]: DEBUG nova.network.neutron [-] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 827.524727] env[63355]: DEBUG nova.network.neutron [-] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.613960] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.846s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.614710] env[63355]: ERROR nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f4801f4b-2b95-41b6-a74c-7c95d83c7152, please check neutron logs for more information. [ 827.614710] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Traceback (most recent call last): [ 827.614710] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 827.614710] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] self.driver.spawn(context, instance, image_meta, [ 827.614710] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 827.614710] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 827.614710] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 827.614710] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] vm_ref = self.build_virtual_machine(instance, [ 827.614710] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 827.614710] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 827.614710] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] for vif in network_info: [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] return self._sync_wrapper(fn, *args, **kwargs) [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] self.wait() [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] self[:] = self._gt.wait() [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] return self._exit_event.wait() [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] result = hub.switch() [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 827.615063] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] return self.greenlet.switch() [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] result = function(*args, **kwargs) [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] return func(*args, **kwargs) [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] raise e [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] nwinfo = self.network_api.allocate_for_instance( [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] created_port_ids = self._update_ports_for_instance( [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] with excutils.save_and_reraise_exception(): [ 827.615371] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] self.force_reraise() [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] raise self.value [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] updated_port = self._update_port( [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] _ensure_no_port_binding_failure(port) [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] raise exception.PortBindingFailed(port_id=port['id']) [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] nova.exception.PortBindingFailed: Binding failed for port f4801f4b-2b95-41b6-a74c-7c95d83c7152, please check neutron logs for more information. [ 827.615686] env[63355]: ERROR nova.compute.manager [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] [ 827.616030] env[63355]: DEBUG nova.compute.utils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Binding failed for port f4801f4b-2b95-41b6-a74c-7c95d83c7152, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 827.617021] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.087s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.618840] env[63355]: INFO nova.compute.claims [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.621591] env[63355]: DEBUG nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Build of instance 7816d808-c9dd-403f-b993-c6d9dc458c7b was re-scheduled: Binding failed for port f4801f4b-2b95-41b6-a74c-7c95d83c7152, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 827.622013] env[63355]: DEBUG nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 827.622250] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Acquiring lock "refresh_cache-7816d808-c9dd-403f-b993-c6d9dc458c7b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.622398] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Acquired lock "refresh_cache-7816d808-c9dd-403f-b993-c6d9dc458c7b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.622557] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.893981] env[63355]: INFO nova.compute.manager [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 795b61f6-a0f3-4aab-94b3-2540ccbb7312] Took 1.02 seconds to deallocate network for instance. [ 828.027344] env[63355]: DEBUG nova.network.neutron [-] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.145458] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.221076] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.529516] env[63355]: INFO nova.compute.manager [-] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Took 1.02 seconds to deallocate network for instance. [ 828.531990] env[63355]: DEBUG nova.compute.claims [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 828.532191] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.723745] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Releasing lock "refresh_cache-7816d808-c9dd-403f-b993-c6d9dc458c7b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.724997] env[63355]: DEBUG nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 828.724997] env[63355]: DEBUG nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 828.724997] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 828.741373] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.921392] env[63355]: INFO nova.scheduler.client.report [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Deleted allocations for instance 795b61f6-a0f3-4aab-94b3-2540ccbb7312 [ 829.037846] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ceeb2f-2086-41d2-a47b-71c73b698a07 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.048487] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba81ebae-135a-4d01-a918-68c5131104d6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.099885] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6abe08a-a6ad-4899-b004-fb8bfdf0bd40 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.111083] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08fb104f-175b-4c61-8ad6-356bffa89ed8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.130753] env[63355]: DEBUG nova.compute.provider_tree [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.244349] env[63355]: DEBUG nova.network.neutron [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.433970] env[63355]: DEBUG oslo_concurrency.lockutils [None req-52d91091-c9ad-401b-9d4a-1fa4edd5f07b tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "795b61f6-a0f3-4aab-94b3-2540ccbb7312" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 174.858s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.635506] env[63355]: DEBUG nova.scheduler.client.report [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.750236] env[63355]: INFO nova.compute.manager [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] [instance: 7816d808-c9dd-403f-b993-c6d9dc458c7b] Took 1.03 seconds to deallocate network for instance. [ 829.939292] env[63355]: DEBUG nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 830.140953] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.141130] env[63355]: DEBUG nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.143647] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.649s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.145062] env[63355]: INFO nova.compute.claims [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.459588] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.649746] env[63355]: DEBUG nova.compute.utils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.651156] env[63355]: DEBUG nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.651322] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 830.694118] env[63355]: DEBUG nova.policy [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd32aea4219944964ba3fa6eed5d70ea7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58271233c0bc4384acb7cd581d3bed59', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.780113] env[63355]: INFO nova.scheduler.client.report [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Deleted allocations for instance 7816d808-c9dd-403f-b993-c6d9dc458c7b [ 830.971731] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Successfully created port: 33d8e84f-04c9-4513-961a-9deff4a565dd {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.155992] env[63355]: DEBUG nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.217907] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "dc56b350-cee7-49c4-9712-8c8c29146ff2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.219210] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "dc56b350-cee7-49c4-9712-8c8c29146ff2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.250610] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.250827] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.292317] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e784e423-6c73-4d3f-af1a-4bed9a6e208e tempest-ServerPasswordTestJSON-334568038 tempest-ServerPasswordTestJSON-334568038-project-member] Lock "7816d808-c9dd-403f-b993-c6d9dc458c7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 167.103s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.492284] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557bf473-53b6-49af-8d3c-2ea31ebbc494 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.500695] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67b21af-aa48-43c1-a3a8-619817f08375 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.539272] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593b36be-bcc8-4790-b9a4-8ade50035611 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.548785] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bad6da-b381-4fbc-b776-707a90eee325 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.563076] env[63355]: DEBUG nova.compute.provider_tree [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.731096] env[63355]: DEBUG nova.compute.manager [req-69e98094-00f6-460c-bacf-745004ad4575 req-2eccf737-1239-4857-b511-7e5abd629255 service nova] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Received event network-changed-33d8e84f-04c9-4513-961a-9deff4a565dd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.731096] env[63355]: DEBUG nova.compute.manager [req-69e98094-00f6-460c-bacf-745004ad4575 req-2eccf737-1239-4857-b511-7e5abd629255 service nova] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Refreshing instance network info cache due to event network-changed-33d8e84f-04c9-4513-961a-9deff4a565dd. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.731096] env[63355]: DEBUG oslo_concurrency.lockutils [req-69e98094-00f6-460c-bacf-745004ad4575 req-2eccf737-1239-4857-b511-7e5abd629255 service nova] Acquiring lock "refresh_cache-dd80c3be-ec92-4a67-a373-3e2276752feb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.731096] env[63355]: DEBUG oslo_concurrency.lockutils [req-69e98094-00f6-460c-bacf-745004ad4575 req-2eccf737-1239-4857-b511-7e5abd629255 service nova] Acquired lock "refresh_cache-dd80c3be-ec92-4a67-a373-3e2276752feb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.731096] env[63355]: DEBUG nova.network.neutron [req-69e98094-00f6-460c-bacf-745004ad4575 req-2eccf737-1239-4857-b511-7e5abd629255 service nova] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Refreshing network info cache for port 33d8e84f-04c9-4513-961a-9deff4a565dd {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.792528] env[63355]: DEBUG nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.858510] env[63355]: ERROR nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 33d8e84f-04c9-4513-961a-9deff4a565dd, please check neutron logs for more information. [ 831.858510] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 831.858510] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 831.858510] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 831.858510] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 831.858510] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 831.858510] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 831.858510] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 831.858510] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 831.858510] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 831.858510] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 831.858510] env[63355]: ERROR nova.compute.manager raise self.value [ 831.858510] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 831.858510] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 831.858510] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 831.858510] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 831.858864] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 831.858864] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 831.858864] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 33d8e84f-04c9-4513-961a-9deff4a565dd, please check neutron logs for more information. [ 831.858864] env[63355]: ERROR nova.compute.manager [ 831.858864] env[63355]: Traceback (most recent call last): [ 831.858864] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 831.858864] env[63355]: listener.cb(fileno) [ 831.858864] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 831.858864] env[63355]: result = function(*args, **kwargs) [ 831.858864] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 831.858864] env[63355]: return func(*args, **kwargs) [ 831.858864] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 831.858864] env[63355]: raise e [ 831.858864] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 831.858864] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 831.858864] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 831.858864] env[63355]: created_port_ids = self._update_ports_for_instance( [ 831.858864] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 831.858864] env[63355]: with excutils.save_and_reraise_exception(): [ 831.858864] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 831.858864] env[63355]: self.force_reraise() [ 831.858864] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 831.858864] env[63355]: raise self.value [ 831.858864] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 831.858864] env[63355]: updated_port = self._update_port( [ 831.858864] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 831.858864] env[63355]: _ensure_no_port_binding_failure(port) [ 831.858864] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 831.858864] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 831.859502] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 33d8e84f-04c9-4513-961a-9deff4a565dd, please check neutron logs for more information. [ 831.859502] env[63355]: Removing descriptor: 17 [ 832.066230] env[63355]: DEBUG nova.scheduler.client.report [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.171108] env[63355]: DEBUG nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.202090] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.202090] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.202090] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.202090] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.202282] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.202282] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.202681] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.202979] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.203312] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.203598] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.203884] env[63355]: DEBUG nova.virt.hardware [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.204840] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff90f7cb-eed4-4e5f-8869-b5fea3381b47 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.214896] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe14c504-b6a4-424b-a79d-414da7127e90 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.231513] env[63355]: ERROR nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 33d8e84f-04c9-4513-961a-9deff4a565dd, please check neutron logs for more information. [ 832.231513] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Traceback (most recent call last): [ 832.231513] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 832.231513] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] yield resources [ 832.231513] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 832.231513] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] self.driver.spawn(context, instance, image_meta, [ 832.231513] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 832.231513] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 832.231513] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 832.231513] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] vm_ref = self.build_virtual_machine(instance, [ 832.231513] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] vif_infos = vmwarevif.get_vif_info(self._session, [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] for vif in network_info: [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] return self._sync_wrapper(fn, *args, **kwargs) [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] self.wait() [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] self[:] = self._gt.wait() [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] return self._exit_event.wait() [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 832.233279] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] current.throw(*self._exc) [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] result = function(*args, **kwargs) [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] return func(*args, **kwargs) [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] raise e [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] nwinfo = self.network_api.allocate_for_instance( [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] created_port_ids = self._update_ports_for_instance( [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] with excutils.save_and_reraise_exception(): [ 832.233695] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] self.force_reraise() [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] raise self.value [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] updated_port = self._update_port( [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] _ensure_no_port_binding_failure(port) [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] raise exception.PortBindingFailed(port_id=port['id']) [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] nova.exception.PortBindingFailed: Binding failed for port 33d8e84f-04c9-4513-961a-9deff4a565dd, please check neutron logs for more information. [ 832.234127] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] [ 832.234127] env[63355]: INFO nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Terminating instance [ 832.237297] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquiring lock "refresh_cache-dd80c3be-ec92-4a67-a373-3e2276752feb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.254025] env[63355]: DEBUG nova.network.neutron [req-69e98094-00f6-460c-bacf-745004ad4575 req-2eccf737-1239-4857-b511-7e5abd629255 service nova] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.313036] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.337831] env[63355]: DEBUG nova.network.neutron [req-69e98094-00f6-460c-bacf-745004ad4575 req-2eccf737-1239-4857-b511-7e5abd629255 service nova] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.573554] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.574085] env[63355]: DEBUG nova.compute.manager [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.576905] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.091s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.839687] env[63355]: DEBUG oslo_concurrency.lockutils [req-69e98094-00f6-460c-bacf-745004ad4575 req-2eccf737-1239-4857-b511-7e5abd629255 service nova] Releasing lock "refresh_cache-dd80c3be-ec92-4a67-a373-3e2276752feb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.840190] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquired lock "refresh_cache-dd80c3be-ec92-4a67-a373-3e2276752feb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.841193] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.083254] env[63355]: DEBUG nova.compute.utils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.089034] env[63355]: DEBUG nova.compute.manager [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Not allocating networking since 'none' was specified. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 833.362068] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.399120] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3815e171-bf68-4067-aa64-1585c3101800 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.406655] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3814de32-f9c1-4e5a-ba17-ce1ee9176871 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.439575] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101b156d-21d6-4c52-9af0-141a20d3f243 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.447744] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f537533-be22-4fe5-8736-a0964726f4c7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.461706] env[63355]: DEBUG nova.compute.provider_tree [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.473065] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.590370] env[63355]: DEBUG nova.compute.manager [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.759497] env[63355]: DEBUG nova.compute.manager [req-91ab821b-e06d-430a-8506-4a19c48110f8 req-9e3406e3-d78f-44eb-b3e1-87c85d40008f service nova] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Received event network-vif-deleted-33d8e84f-04c9-4513-961a-9deff4a565dd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.967896] env[63355]: DEBUG nova.scheduler.client.report [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.975523] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Releasing lock "refresh_cache-dd80c3be-ec92-4a67-a373-3e2276752feb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.976055] env[63355]: DEBUG nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 833.976281] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 833.976570] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ce24fda4-bb50-4e72-aa29-937132688e61 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.987118] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55d5f57-217d-4dcf-9828-4c613e14cc58 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.010341] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dd80c3be-ec92-4a67-a373-3e2276752feb could not be found. [ 834.010620] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.010812] env[63355]: INFO nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 834.011062] env[63355]: DEBUG oslo.service.loopingcall [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.011532] env[63355]: DEBUG nova.compute.manager [-] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.011631] env[63355]: DEBUG nova.network.neutron [-] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.027404] env[63355]: DEBUG nova.network.neutron [-] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.473162] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.896s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.473811] env[63355]: ERROR nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3d179281-1687-4048-bd8a-8e574b2dd566, please check neutron logs for more information. [ 834.473811] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Traceback (most recent call last): [ 834.473811] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 834.473811] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] self.driver.spawn(context, instance, image_meta, [ 834.473811] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 834.473811] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] self._vmops.spawn(context, instance, image_meta, injected_files, [ 834.473811] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 834.473811] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] vm_ref = self.build_virtual_machine(instance, [ 834.473811] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 834.473811] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] vif_infos = vmwarevif.get_vif_info(self._session, [ 834.473811] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] for vif in network_info: [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] return self._sync_wrapper(fn, *args, **kwargs) [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] self.wait() [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] self[:] = self._gt.wait() [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] return self._exit_event.wait() [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] result = hub.switch() [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 834.474126] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] return self.greenlet.switch() [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] result = function(*args, **kwargs) [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] return func(*args, **kwargs) [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] raise e [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] nwinfo = self.network_api.allocate_for_instance( [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] created_port_ids = self._update_ports_for_instance( [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] with excutils.save_and_reraise_exception(): [ 834.474394] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] self.force_reraise() [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] raise self.value [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] updated_port = self._update_port( [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] _ensure_no_port_binding_failure(port) [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] raise exception.PortBindingFailed(port_id=port['id']) [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] nova.exception.PortBindingFailed: Binding failed for port 3d179281-1687-4048-bd8a-8e574b2dd566, please check neutron logs for more information. [ 834.474661] env[63355]: ERROR nova.compute.manager [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] [ 834.474914] env[63355]: DEBUG nova.compute.utils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Binding failed for port 3d179281-1687-4048-bd8a-8e574b2dd566, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 834.475898] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.511s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.477486] env[63355]: INFO nova.compute.claims [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.480225] env[63355]: DEBUG nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Build of instance a1e1e0ea-debd-4529-b175-4bdf44e4de31 was re-scheduled: Binding failed for port 3d179281-1687-4048-bd8a-8e574b2dd566, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 834.480655] env[63355]: DEBUG nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 834.480873] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquiring lock "refresh_cache-a1e1e0ea-debd-4529-b175-4bdf44e4de31" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.481037] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquired lock "refresh_cache-a1e1e0ea-debd-4529-b175-4bdf44e4de31" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.481206] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.530454] env[63355]: DEBUG nova.network.neutron [-] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.601173] env[63355]: DEBUG nova.compute.manager [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.626395] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.626636] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.626792] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.626969] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.627134] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.627281] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.627481] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.627640] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.627799] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.627959] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.628168] env[63355]: DEBUG nova.virt.hardware [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.629027] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b48e215-ed2c-469d-b502-9aba2b95cae3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.637440] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1a4385-b568-48ec-8f1d-02750ba6f123 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.651147] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.656838] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Creating folder: Project (165a8af01a094c1d837b68e92a30025e). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 834.656973] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ba7deb21-a309-4d76-8106-fca58dc679a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.670350] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Created folder: Project (165a8af01a094c1d837b68e92a30025e) in parent group-v287607. [ 834.670534] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Creating folder: Instances. Parent ref: group-v287632. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 834.670766] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dae39b5c-2b13-4aa1-8606-9e2f0962a6eb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.680924] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Created folder: Instances in parent group-v287632. [ 834.681174] env[63355]: DEBUG oslo.service.loopingcall [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.681396] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 834.681614] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7213ba4-3f1e-4d09-ae8c-be195b8dceb3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.698403] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.698403] env[63355]: value = "task-1349565" [ 834.698403] env[63355]: _type = "Task" [ 834.698403] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.705909] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349565, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.001698] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.032701] env[63355]: INFO nova.compute.manager [-] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Took 1.02 seconds to deallocate network for instance. [ 835.034803] env[63355]: DEBUG nova.compute.claims [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 835.034929] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.071756] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.209045] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349565, 'name': CreateVM_Task, 'duration_secs': 0.254249} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.209536] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 835.209971] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.210148] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.210466] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.210707] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14c7a390-d168-4c0c-9f21-03520b01cdb6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.215495] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 835.215495] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e9606a-0f36-d0a2-4bfa-a2e120a0d6ea" [ 835.215495] env[63355]: _type = "Task" [ 835.215495] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.223648] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e9606a-0f36-d0a2-4bfa-a2e120a0d6ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.574058] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Releasing lock "refresh_cache-a1e1e0ea-debd-4529-b175-4bdf44e4de31" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.574319] env[63355]: DEBUG nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 835.574502] env[63355]: DEBUG nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 835.574670] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 835.592197] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.726088] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e9606a-0f36-d0a2-4bfa-a2e120a0d6ea, 'name': SearchDatastore_Task, 'duration_secs': 0.011963} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.728329] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.728562] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.728790] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.728934] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.729123] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.730145] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ddc45a26-c850-41db-9f70-7a8b5ea497b8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.741079] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.741323] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.742134] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ed59c8a-28a7-469e-a47f-fc39aeb94af0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.749920] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 835.749920] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a2e94d-37fd-895f-19ec-6b65ea4a9a5c" [ 835.749920] env[63355]: _type = "Task" [ 835.749920] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.760439] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a2e94d-37fd-895f-19ec-6b65ea4a9a5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.908414] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ecaa03-e35a-4918-a88c-79f5da7cc5db {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.916905] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89bad99-21f8-4e30-8266-9e4c1a028f83 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.948319] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1fc25a-b4be-45eb-9c98-c6e314414190 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.957757] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cfa9a05-9675-4c67-b592-1d6fef1f6835 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.973158] env[63355]: DEBUG nova.compute.provider_tree [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.095373] env[63355]: DEBUG nova.network.neutron [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.263471] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a2e94d-37fd-895f-19ec-6b65ea4a9a5c, 'name': SearchDatastore_Task, 'duration_secs': 0.009572} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.264261] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1982ebc8-6e49-4b7e-bbc4-f2c968ae90bd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.269869] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 836.269869] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529d5e9e-ac5f-250b-513f-b8ed3e7b026c" [ 836.269869] env[63355]: _type = "Task" [ 836.269869] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.277843] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529d5e9e-ac5f-250b-513f-b8ed3e7b026c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.475927] env[63355]: DEBUG nova.scheduler.client.report [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.598749] env[63355]: INFO nova.compute.manager [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: a1e1e0ea-debd-4529-b175-4bdf44e4de31] Took 1.02 seconds to deallocate network for instance. [ 836.780193] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529d5e9e-ac5f-250b-513f-b8ed3e7b026c, 'name': SearchDatastore_Task, 'duration_secs': 0.009814} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.780465] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.780719] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 836.780969] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08d19a41-7196-457e-9377-cca13a484b87 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.788955] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 836.788955] env[63355]: value = "task-1349566" [ 836.788955] env[63355]: _type = "Task" [ 836.788955] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.798050] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349566, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.980800] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.505s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.981531] env[63355]: DEBUG nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.984940] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.143s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.300115] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349566, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468408} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.300362] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.301712] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.301712] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-17e963ea-6e48-41a3-abbd-15a69999b34f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.309105] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 837.309105] env[63355]: value = "task-1349567" [ 837.309105] env[63355]: _type = "Task" [ 837.309105] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.318626] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349567, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.490380] env[63355]: DEBUG nova.compute.utils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.494989] env[63355]: DEBUG nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.495282] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 837.517443] env[63355]: DEBUG nova.scheduler.client.report [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Refreshing inventories for resource provider 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 837.530405] env[63355]: DEBUG nova.scheduler.client.report [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Updating ProviderTree inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 837.530568] env[63355]: DEBUG nova.compute.provider_tree [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 837.539582] env[63355]: DEBUG nova.policy [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fdbd89aac89e4d58bbfd5b549d42a27f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '744ce7dc0ab4463fa461cbac87e51904', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.541643] env[63355]: DEBUG nova.scheduler.client.report [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Refreshing aggregate associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, aggregates: None {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 837.557976] env[63355]: DEBUG nova.scheduler.client.report [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Refreshing trait associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 837.624518] env[63355]: INFO nova.scheduler.client.report [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Deleted allocations for instance a1e1e0ea-debd-4529-b175-4bdf44e4de31 [ 837.819237] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349567, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072324} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.822257] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.822890] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cb989d-affb-4291-90ff-11618546449d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.843831] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.844693] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Successfully created port: e7c1c9cf-6026-408e-8353-37aeb78418d7 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.848384] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50c870e4-60c2-43ec-811c-605b53b0083a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.869724] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 837.869724] env[63355]: value = "task-1349568" [ 837.869724] env[63355]: _type = "Task" [ 837.869724] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.879117] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349568, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.942665] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0620d603-0c78-4d77-87c0-be07d260faba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.953969] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c176226-3027-4398-ab45-1bcf942bc9cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.990185] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04880450-a9cc-49b4-ba41-6c132d954c1b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.998916] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd0b351-02a3-437e-b9ae-deb86230c7e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.003225] env[63355]: DEBUG nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 838.019435] env[63355]: DEBUG nova.compute.provider_tree [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.093169] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Acquiring lock "ecbacdb3-516e-43c3-96e3-4961b76565ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.093454] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "ecbacdb3-516e-43c3-96e3-4961b76565ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.134962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6c617fdc-5036-49ee-8201-8ec45fba3cbe tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "a1e1e0ea-debd-4529-b175-4bdf44e4de31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 169.508s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.379843] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349568, 'name': ReconfigVM_Task, 'duration_secs': 0.286585} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.380155] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.380751] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10939fb7-7858-4628-9d68-01da03d81ebe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.388371] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 838.388371] env[63355]: value = "task-1349569" [ 838.388371] env[63355]: _type = "Task" [ 838.388371] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.397115] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349569, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.525191] env[63355]: DEBUG nova.scheduler.client.report [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.557749] env[63355]: DEBUG nova.compute.manager [req-de7d50de-5796-400c-8342-0965272ef308 req-bc709e75-19e3-47a8-a7d4-58724ebd6f18 service nova] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Received event network-changed-e7c1c9cf-6026-408e-8353-37aeb78418d7 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.557912] env[63355]: DEBUG nova.compute.manager [req-de7d50de-5796-400c-8342-0965272ef308 req-bc709e75-19e3-47a8-a7d4-58724ebd6f18 service nova] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Refreshing instance network info cache due to event network-changed-e7c1c9cf-6026-408e-8353-37aeb78418d7. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.558152] env[63355]: DEBUG oslo_concurrency.lockutils [req-de7d50de-5796-400c-8342-0965272ef308 req-bc709e75-19e3-47a8-a7d4-58724ebd6f18 service nova] Acquiring lock "refresh_cache-8a44bcfb-406c-45f6-b027-86d8980bc5d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.558290] env[63355]: DEBUG oslo_concurrency.lockutils [req-de7d50de-5796-400c-8342-0965272ef308 req-bc709e75-19e3-47a8-a7d4-58724ebd6f18 service nova] Acquired lock "refresh_cache-8a44bcfb-406c-45f6-b027-86d8980bc5d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.558442] env[63355]: DEBUG nova.network.neutron [req-de7d50de-5796-400c-8342-0965272ef308 req-bc709e75-19e3-47a8-a7d4-58724ebd6f18 service nova] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Refreshing network info cache for port e7c1c9cf-6026-408e-8353-37aeb78418d7 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.639517] env[63355]: DEBUG nova.compute.manager [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.714476] env[63355]: ERROR nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e7c1c9cf-6026-408e-8353-37aeb78418d7, please check neutron logs for more information. [ 838.714476] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 838.714476] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 838.714476] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 838.714476] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 838.714476] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 838.714476] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 838.714476] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 838.714476] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 838.714476] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 838.714476] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 838.714476] env[63355]: ERROR nova.compute.manager raise self.value [ 838.714476] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 838.714476] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 838.714476] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 838.714476] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 838.715043] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 838.715043] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 838.715043] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e7c1c9cf-6026-408e-8353-37aeb78418d7, please check neutron logs for more information. [ 838.715043] env[63355]: ERROR nova.compute.manager [ 838.715043] env[63355]: Traceback (most recent call last): [ 838.715043] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 838.715043] env[63355]: listener.cb(fileno) [ 838.715043] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 838.715043] env[63355]: result = function(*args, **kwargs) [ 838.715043] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 838.715043] env[63355]: return func(*args, **kwargs) [ 838.715043] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 838.715043] env[63355]: raise e [ 838.715043] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 838.715043] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 838.715043] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 838.715043] env[63355]: created_port_ids = self._update_ports_for_instance( [ 838.715043] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 838.715043] env[63355]: with excutils.save_and_reraise_exception(): [ 838.715043] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 838.715043] env[63355]: self.force_reraise() [ 838.715043] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 838.715043] env[63355]: raise self.value [ 838.715043] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 838.715043] env[63355]: updated_port = self._update_port( [ 838.715043] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 838.715043] env[63355]: _ensure_no_port_binding_failure(port) [ 838.715043] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 838.715043] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 838.715650] env[63355]: nova.exception.PortBindingFailed: Binding failed for port e7c1c9cf-6026-408e-8353-37aeb78418d7, please check neutron logs for more information. [ 838.715650] env[63355]: Removing descriptor: 17 [ 838.898932] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349569, 'name': Rename_Task, 'duration_secs': 0.152566} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.899238] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 838.900041] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1e34132-2378-45e3-ad1b-d5908ac8fd4b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.906639] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 838.906639] env[63355]: value = "task-1349570" [ 838.906639] env[63355]: _type = "Task" [ 838.906639] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.915723] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.030457] env[63355]: DEBUG nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 839.033044] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.048s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.033682] env[63355]: ERROR nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7113633c-13bf-4d05-a697-7df302bda34e, please check neutron logs for more information. [ 839.033682] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Traceback (most recent call last): [ 839.033682] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 839.033682] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] self.driver.spawn(context, instance, image_meta, [ 839.033682] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 839.033682] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 839.033682] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 839.033682] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] vm_ref = self.build_virtual_machine(instance, [ 839.033682] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 839.033682] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] vif_infos = vmwarevif.get_vif_info(self._session, [ 839.033682] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] for vif in network_info: [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] return self._sync_wrapper(fn, *args, **kwargs) [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] self.wait() [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] self[:] = self._gt.wait() [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] return self._exit_event.wait() [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] current.throw(*self._exc) [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 839.033940] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] result = function(*args, **kwargs) [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] return func(*args, **kwargs) [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] raise e [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] nwinfo = self.network_api.allocate_for_instance( [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] created_port_ids = self._update_ports_for_instance( [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] with excutils.save_and_reraise_exception(): [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] self.force_reraise() [ 839.034231] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 839.034560] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] raise self.value [ 839.034560] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 839.034560] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] updated_port = self._update_port( [ 839.034560] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 839.034560] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] _ensure_no_port_binding_failure(port) [ 839.034560] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 839.034560] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] raise exception.PortBindingFailed(port_id=port['id']) [ 839.034560] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] nova.exception.PortBindingFailed: Binding failed for port 7113633c-13bf-4d05-a697-7df302bda34e, please check neutron logs for more information. [ 839.034560] env[63355]: ERROR nova.compute.manager [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] [ 839.034560] env[63355]: DEBUG nova.compute.utils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Binding failed for port 7113633c-13bf-4d05-a697-7df302bda34e, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 839.035766] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.200s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.038245] env[63355]: DEBUG nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Build of instance 21e17d32-40db-4dc1-9e73-fa6d35a4c06a was re-scheduled: Binding failed for port 7113633c-13bf-4d05-a697-7df302bda34e, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 839.038684] env[63355]: DEBUG nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 839.038905] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "refresh_cache-21e17d32-40db-4dc1-9e73-fa6d35a4c06a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.039061] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquired lock "refresh_cache-21e17d32-40db-4dc1-9e73-fa6d35a4c06a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.039224] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 839.059496] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.059813] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.060011] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.060213] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.060388] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.060567] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.060889] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.061195] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.061425] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.061635] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.062110] env[63355]: DEBUG nova.virt.hardware [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.065156] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94f9a46-7b3b-436d-a075-a2bad33f231d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.075275] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d345456-4ed9-4905-b1d6-81c2070fa208 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.083476] env[63355]: DEBUG nova.network.neutron [req-de7d50de-5796-400c-8342-0965272ef308 req-bc709e75-19e3-47a8-a7d4-58724ebd6f18 service nova] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 839.094885] env[63355]: ERROR nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e7c1c9cf-6026-408e-8353-37aeb78418d7, please check neutron logs for more information. [ 839.094885] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Traceback (most recent call last): [ 839.094885] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 839.094885] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] yield resources [ 839.094885] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 839.094885] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] self.driver.spawn(context, instance, image_meta, [ 839.094885] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 839.094885] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 839.094885] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 839.094885] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] vm_ref = self.build_virtual_machine(instance, [ 839.094885] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] for vif in network_info: [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] return self._sync_wrapper(fn, *args, **kwargs) [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] self.wait() [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] self[:] = self._gt.wait() [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] return self._exit_event.wait() [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 839.095465] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] current.throw(*self._exc) [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] result = function(*args, **kwargs) [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] return func(*args, **kwargs) [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] raise e [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] nwinfo = self.network_api.allocate_for_instance( [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] created_port_ids = self._update_ports_for_instance( [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] with excutils.save_and_reraise_exception(): [ 839.095844] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] self.force_reraise() [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] raise self.value [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] updated_port = self._update_port( [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] _ensure_no_port_binding_failure(port) [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] raise exception.PortBindingFailed(port_id=port['id']) [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] nova.exception.PortBindingFailed: Binding failed for port e7c1c9cf-6026-408e-8353-37aeb78418d7, please check neutron logs for more information. [ 839.096210] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] [ 839.096210] env[63355]: INFO nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Terminating instance [ 839.097125] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "refresh_cache-8a44bcfb-406c-45f6-b027-86d8980bc5d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.157853] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.171445] env[63355]: DEBUG nova.network.neutron [req-de7d50de-5796-400c-8342-0965272ef308 req-bc709e75-19e3-47a8-a7d4-58724ebd6f18 service nova] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.416902] env[63355]: DEBUG oslo_vmware.api [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349570, 'name': PowerOnVM_Task, 'duration_secs': 0.425104} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.417211] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.417398] env[63355]: INFO nova.compute.manager [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Took 4.82 seconds to spawn the instance on the hypervisor. [ 839.417576] env[63355]: DEBUG nova.compute.manager [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.418337] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0825f5e8-283f-419e-8d88-3aa8599e96e9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.560544] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 839.631145] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.675938] env[63355]: DEBUG oslo_concurrency.lockutils [req-de7d50de-5796-400c-8342-0965272ef308 req-bc709e75-19e3-47a8-a7d4-58724ebd6f18 service nova] Releasing lock "refresh_cache-8a44bcfb-406c-45f6-b027-86d8980bc5d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.675938] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired lock "refresh_cache-8a44bcfb-406c-45f6-b027-86d8980bc5d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.675938] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 839.939775] env[63355]: INFO nova.compute.manager [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Took 25.46 seconds to build instance. [ 840.134873] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Releasing lock "refresh_cache-21e17d32-40db-4dc1-9e73-fa6d35a4c06a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.135243] env[63355]: DEBUG nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 840.135434] env[63355]: DEBUG nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.135597] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 840.151526] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.193063] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.273476] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.383099] env[63355]: INFO nova.compute.manager [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Rebuilding instance [ 840.422869] env[63355]: DEBUG nova.compute.manager [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.423736] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390841d5-8898-42d8-a03a-e0b30f808e6f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.442248] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12caff62-7ccf-444e-8a29-7accf525537d tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "9989c129-07d3-4af9-9a86-1e2746d6ac70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.878s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.571232] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 21e17d32-40db-4dc1-9e73-fa6d35a4c06a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 840.571405] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7d79c9cf-66a9-4117-b090-0ab0676c3114 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.571532] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 4ded6053-9afc-444e-bdfc-244265ed19a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.571651] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance dd80c3be-ec92-4a67-a373-3e2276752feb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.571800] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 9989c129-07d3-4af9-9a86-1e2746d6ac70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.571865] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8a44bcfb-406c-45f6-b027-86d8980bc5d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.587128] env[63355]: DEBUG nova.compute.manager [req-6c07bec4-fc1a-4abd-aa6c-657d2fb5092e req-04e1e9f1-c92b-4de3-afce-fa09e7cc0db2 service nova] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Received event network-vif-deleted-e7c1c9cf-6026-408e-8353-37aeb78418d7 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.653467] env[63355]: DEBUG nova.network.neutron [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.776212] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Releasing lock "refresh_cache-8a44bcfb-406c-45f6-b027-86d8980bc5d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.776697] env[63355]: DEBUG nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.776892] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.777224] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44114b49-0e9e-45ef-bdb6-adac6160b0dd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.787474] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f7b9ca-64cb-444b-a728-cf3ce3a9c6ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.810316] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8a44bcfb-406c-45f6-b027-86d8980bc5d1 could not be found. [ 840.810497] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 840.810677] env[63355]: INFO nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 840.810915] env[63355]: DEBUG oslo.service.loopingcall [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.811173] env[63355]: DEBUG nova.compute.manager [-] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.811271] env[63355]: DEBUG nova.network.neutron [-] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 840.826328] env[63355]: DEBUG nova.network.neutron [-] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.935103] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.935419] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b43921c-41ce-44e1-81da-728149f088f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.945176] env[63355]: DEBUG nova.compute.manager [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 840.948517] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 840.948517] env[63355]: value = "task-1349571" [ 840.948517] env[63355]: _type = "Task" [ 840.948517] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.957467] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.076642] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance e49ba26c-486c-43aa-9595-231e7f31a522 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 841.156633] env[63355]: INFO nova.compute.manager [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: 21e17d32-40db-4dc1-9e73-fa6d35a4c06a] Took 1.02 seconds to deallocate network for instance. [ 841.329278] env[63355]: DEBUG nova.network.neutron [-] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.463744] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349571, 'name': PowerOffVM_Task, 'duration_secs': 0.187658} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.464025] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 841.464249] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 841.464974] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c827a5-b116-4648-845b-6bac30ae13a9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.468141] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.473279] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 841.473481] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b0b900d-b709-4e21-b856-8885b61777ca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.500723] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.500947] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.501152] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Deleting the datastore file [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.501672] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b8684803-7f3a-4262-a822-eb0915aa4e20 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.508889] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 841.508889] env[63355]: value = "task-1349573" [ 841.508889] env[63355]: _type = "Task" [ 841.508889] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.516821] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349573, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.580022] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7dd28cb7-d249-4b8e-beb3-c959f15e912f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 841.830733] env[63355]: INFO nova.compute.manager [-] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Took 1.02 seconds to deallocate network for instance. [ 841.833122] env[63355]: DEBUG nova.compute.claims [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 841.833304] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.018573] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349573, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106388} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.018838] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.019046] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 842.019234] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 842.083951] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 947a33d3-88c1-45ae-805c-eda2bc7626f1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 842.183511] env[63355]: INFO nova.scheduler.client.report [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Deleted allocations for instance 21e17d32-40db-4dc1-9e73-fa6d35a4c06a [ 842.587151] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 3d918cab-6ed6-4a37-a024-28e3db1b779c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 842.693039] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01d21293-1981-4c82-a584-38140b4efe68 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "21e17d32-40db-4dc1-9e73-fa6d35a4c06a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 173.982s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.056145] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.056279] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.056435] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.056620] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.056764] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.056909] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.057127] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.057287] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.057451] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.057610] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.057780] env[63355]: DEBUG nova.virt.hardware [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.058649] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1aed740-7068-4673-819c-98cdd84515d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.067013] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f48526-4991-4f22-ae57-577c094db69a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.080789] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.086626] env[63355]: DEBUG oslo.service.loopingcall [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.086876] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.087900] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c37579b2-ad49-4dbc-85f7-17be825e4d92 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.099698] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 5591bb55-83d7-4301-a3f9-fde945632344 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 843.107576] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.107576] env[63355]: value = "task-1349574" [ 843.107576] env[63355]: _type = "Task" [ 843.107576] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.116429] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349574, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.196120] env[63355]: DEBUG nova.compute.manager [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.579235] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.579482] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.603215] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 386e847e-967b-4247-9730-cdc5ac251474 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 843.618549] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349574, 'name': CreateVM_Task, 'duration_secs': 0.273666} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.618549] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.618724] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.618759] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.619127] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.619382] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2724f828-190c-427a-b8de-94673b299807 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.624598] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 843.624598] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52684e05-6baf-2089-b561-8fdb0a2c27fb" [ 843.624598] env[63355]: _type = "Task" [ 843.624598] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.633216] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52684e05-6baf-2089-b561-8fdb0a2c27fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.715354] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.106077] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 0b660647-697a-445b-bca4-82ef6975ea75 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 844.135287] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52684e05-6baf-2089-b561-8fdb0a2c27fb, 'name': SearchDatastore_Task, 'duration_secs': 0.009722} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.135736] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.136022] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.136269] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.136418] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.136593] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.136850] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a33d197f-cbc0-4624-b48b-2211bef437e3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.146885] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.146983] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.147723] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f23a11f9-1106-4122-b0c5-adfdfb35c55b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.153063] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 844.153063] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52679d1e-2520-7cee-a2f8-b75be4a3110b" [ 844.153063] env[63355]: _type = "Task" [ 844.153063] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.161498] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52679d1e-2520-7cee-a2f8-b75be4a3110b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.609649] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 844.663531] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52679d1e-2520-7cee-a2f8-b75be4a3110b, 'name': SearchDatastore_Task, 'duration_secs': 0.009009} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.664334] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2a8b349-1aeb-456f-8efb-1cbe36b2cfa7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.669911] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 844.669911] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526f9d95-c7c7-1177-6e0f-ac884f01a29e" [ 844.669911] env[63355]: _type = "Task" [ 844.669911] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.678583] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526f9d95-c7c7-1177-6e0f-ac884f01a29e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.112875] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance f5e62ce1-40b7-4648-a4a6-068ff06eaf9b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 845.181576] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526f9d95-c7c7-1177-6e0f-ac884f01a29e, 'name': SearchDatastore_Task, 'duration_secs': 0.008959} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.181853] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.182122] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.182376] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d364b49-b7c2-4d27-b24a-c7ffd6ec1a36 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.190860] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 845.190860] env[63355]: value = "task-1349575" [ 845.190860] env[63355]: _type = "Task" [ 845.190860] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.199665] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349575, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.616175] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8e3ec9d3-bc22-4e39-ad7c-93268dd59020 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 845.701573] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349575, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448466} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.701837] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.702102] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.702389] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-46b38f01-f7f8-411b-8d1c-d1979291a07a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.710314] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 845.710314] env[63355]: value = "task-1349576" [ 845.710314] env[63355]: _type = "Task" [ 845.710314] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.719021] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349576, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.119648] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b6056441-9ee4-484f-a1d2-077546f2c581 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 846.220770] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349576, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062363} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.221064] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.221835] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0402fe26-ff5f-4a3f-9671-74a4c1041981 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.242526] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.242895] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-461e5430-bf52-47e7-9492-c2cb8777f214 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.262370] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 846.262370] env[63355]: value = "task-1349577" [ 846.262370] env[63355]: _type = "Task" [ 846.262370] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.270396] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349577, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.624142] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b4b09b1d-680e-47b8-aa8a-9b3d9167824d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 846.773062] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349577, 'name': ReconfigVM_Task, 'duration_secs': 0.280439} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.773062] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.773062] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4177ae5e-a597-4410-aff6-d3fb98e85110 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.779137] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 846.779137] env[63355]: value = "task-1349578" [ 846.779137] env[63355]: _type = "Task" [ 846.779137] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.786516] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349578, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.127602] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 2539a79e-01c0-4e0c-aa66-8784441c6fda has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 847.288883] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349578, 'name': Rename_Task, 'duration_secs': 0.135692} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.289173] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.289413] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04250866-6919-4a12-8481-b99914eea2c5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.295123] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 847.295123] env[63355]: value = "task-1349579" [ 847.295123] env[63355]: _type = "Task" [ 847.295123] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.302450] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349579, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.631029] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7c139710-d8d9-4cd7-bec0-6e021d3b2e68 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 847.805275] env[63355]: DEBUG oslo_vmware.api [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349579, 'name': PowerOnVM_Task, 'duration_secs': 0.402789} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.805559] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 847.805769] env[63355]: DEBUG nova.compute.manager [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.806558] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fd9a1a-9bea-4521-add6-0e311d7d02b3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.134789] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 74071e73-10fa-4dcb-aa15-91303b2278fb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 848.322225] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.638355] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 849.142025] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance dc56b350-cee7-49c4-9712-8c8c29146ff2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 849.258354] env[63355]: INFO nova.compute.manager [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Rebuilding instance [ 849.299554] env[63355]: DEBUG nova.compute.manager [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.300490] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029a6e56-8980-4a20-b19b-311f2ea5da15 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.645058] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 199bc488-2e5d-4cea-aefb-ddd35ecc7a30 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 849.810861] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 849.811109] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9a51a33-7a5f-459e-bc4f-0f2e375b3a64 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.818900] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for the task: (returnval){ [ 849.818900] env[63355]: value = "task-1349580" [ 849.818900] env[63355]: _type = "Task" [ 849.818900] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.826232] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349580, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.147280] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance ecbacdb3-516e-43c3-96e3-4961b76565ca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 850.147564] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 850.147713] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 850.330273] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349580, 'name': PowerOffVM_Task, 'duration_secs': 0.249947} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.330549] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 850.330779] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.331576] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520ca458-6991-4288-996c-839469d65527 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.340200] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 850.340436] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4324a4c5-827b-44b3-98cd-bf543ccc595d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.368945] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 850.369194] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 850.369375] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Deleting the datastore file [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.369632] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-590d7ce4-5d2b-456d-b221-0dc398393598 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.376933] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for the task: (returnval){ [ 850.376933] env[63355]: value = "task-1349582" [ 850.376933] env[63355]: _type = "Task" [ 850.376933] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.387061] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349582, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.412925] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a99c4a-af7c-46ab-8b03-1fa08bc99fa2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.420222] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55459611-6e3a-46de-befd-ffb34889e839 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.449426] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057e2283-c38a-4c5c-9452-1dfe2c4d6336 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.457314] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53189c3-3d5a-4157-9d05-98b99fd08e72 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.471917] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.889377] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349582, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093159} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.889664] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.889816] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 850.889985] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 850.975109] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.480755] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 851.480934] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.445s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.481219] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.618s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.484037] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.484164] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Cleaning up deleted instances {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 851.921854] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.922385] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.922385] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.922471] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.922564] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.922715] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.922938] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.923234] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.923495] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.923692] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.923887] env[63355]: DEBUG nova.virt.hardware [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.924752] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820f3bdb-a9af-445f-821a-06ebf8d0b736 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.932574] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03dd9378-3de8-4d3c-bfdc-0bfcfe6930d1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.945414] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.950768] env[63355]: DEBUG oslo.service.loopingcall [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.950995] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 851.951208] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9c58eec-a8f9-41bb-b363-b689dcb38e1c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.967467] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.967467] env[63355]: value = "task-1349583" [ 851.967467] env[63355]: _type = "Task" [ 851.967467] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.974652] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349583, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.990314] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] There are 5 instances to clean {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 851.990569] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: e5c80c05-0e48-415a-ac91-a53e1c8819f0] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 852.273736] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377a0981-b4d3-4285-905b-a02a18f83d01 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.281062] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0820009-ebc5-4dd5-9e76-e0f252739c69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.311457] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c64fa5-6c22-4ef9-b0f6-567443e9c2c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.318592] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc32161a-d053-48b0-b0af-ddc250fd23aa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.331065] env[63355]: DEBUG nova.compute.provider_tree [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.477056] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349583, 'name': CreateVM_Task, 'duration_secs': 0.255137} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.477232] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 852.477651] env[63355]: DEBUG oslo_concurrency.lockutils [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.477811] env[63355]: DEBUG oslo_concurrency.lockutils [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.478155] env[63355]: DEBUG oslo_concurrency.lockutils [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 852.478399] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f677c921-29d1-4eae-b532-fbcbbf95db3b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.482845] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for the task: (returnval){ [ 852.482845] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fcbb12-7af0-5ac1-b649-75e844b460ae" [ 852.482845] env[63355]: _type = "Task" [ 852.482845] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.490420] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fcbb12-7af0-5ac1-b649-75e844b460ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.494021] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: b2bf4912-6b88-4ece-95c9-e9fd1cf906df] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 852.834027] env[63355]: DEBUG nova.scheduler.client.report [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.993018] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fcbb12-7af0-5ac1-b649-75e844b460ae, 'name': SearchDatastore_Task, 'duration_secs': 0.01175} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.993347] env[63355]: DEBUG oslo_concurrency.lockutils [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.993577] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.993806] env[63355]: DEBUG oslo_concurrency.lockutils [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.993948] env[63355]: DEBUG oslo_concurrency.lockutils [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.994146] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.994422] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9cdfe1b1-80e8-465d-a8ab-66a0f6cd715f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.996714] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: f9c8ed86-506b-4654-8e7e-cb218605764b] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 853.002268] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.002409] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.003133] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c63c490-a4dd-45f7-960c-e7c1cbbf6ae7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.009019] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for the task: (returnval){ [ 853.009019] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526c222a-38b2-fc2d-5a72-e0d1d86ee020" [ 853.009019] env[63355]: _type = "Task" [ 853.009019] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.014968] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526c222a-38b2-fc2d-5a72-e0d1d86ee020, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.339207] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.858s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.339863] env[63355]: ERROR nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7d88a38c-9dc5-4b16-bcee-211885563c6a, please check neutron logs for more information. [ 853.339863] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Traceback (most recent call last): [ 853.339863] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 853.339863] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] self.driver.spawn(context, instance, image_meta, [ 853.339863] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 853.339863] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] self._vmops.spawn(context, instance, image_meta, injected_files, [ 853.339863] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 853.339863] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] vm_ref = self.build_virtual_machine(instance, [ 853.339863] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 853.339863] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] vif_infos = vmwarevif.get_vif_info(self._session, [ 853.339863] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] for vif in network_info: [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] return self._sync_wrapper(fn, *args, **kwargs) [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] self.wait() [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] self[:] = self._gt.wait() [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] return self._exit_event.wait() [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] result = hub.switch() [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 853.340240] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] return self.greenlet.switch() [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] result = function(*args, **kwargs) [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] return func(*args, **kwargs) [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] raise e [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] nwinfo = self.network_api.allocate_for_instance( [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] created_port_ids = self._update_ports_for_instance( [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] with excutils.save_and_reraise_exception(): [ 853.340583] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] self.force_reraise() [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] raise self.value [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] updated_port = self._update_port( [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] _ensure_no_port_binding_failure(port) [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] raise exception.PortBindingFailed(port_id=port['id']) [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] nova.exception.PortBindingFailed: Binding failed for port 7d88a38c-9dc5-4b16-bcee-211885563c6a, please check neutron logs for more information. [ 853.340902] env[63355]: ERROR nova.compute.manager [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] [ 853.341202] env[63355]: DEBUG nova.compute.utils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Binding failed for port 7d88a38c-9dc5-4b16-bcee-211885563c6a, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 853.341829] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.284s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.343366] env[63355]: INFO nova.compute.claims [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.345935] env[63355]: DEBUG nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Build of instance 7d79c9cf-66a9-4117-b090-0ab0676c3114 was re-scheduled: Binding failed for port 7d88a38c-9dc5-4b16-bcee-211885563c6a, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 853.346408] env[63355]: DEBUG nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 853.346629] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquiring lock "refresh_cache-7d79c9cf-66a9-4117-b090-0ab0676c3114" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.346773] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Acquired lock "refresh_cache-7d79c9cf-66a9-4117-b090-0ab0676c3114" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.346926] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.499928] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: def90353-ae90-4e1e-9fe1-8e5459a7309d] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 853.517415] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526c222a-38b2-fc2d-5a72-e0d1d86ee020, 'name': SearchDatastore_Task, 'duration_secs': 0.01014} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.518149] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b9be669-9b20-4b05-8cd1-a7c689a410ce {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.522904] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for the task: (returnval){ [ 853.522904] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5247922d-f730-f766-ae93-7272a599242f" [ 853.522904] env[63355]: _type = "Task" [ 853.522904] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.530107] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5247922d-f730-f766-ae93-7272a599242f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.870981] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.959796] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.002451] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: abe6726e-7d6a-46e0-9273-de3e681311eb] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 854.033955] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5247922d-f730-f766-ae93-7272a599242f, 'name': SearchDatastore_Task, 'duration_secs': 0.009076} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.034243] env[63355]: DEBUG oslo_concurrency.lockutils [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.034580] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 854.034845] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bcea60f4-1968-4b9a-9ad4-f190349e88e3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.041165] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for the task: (returnval){ [ 854.041165] env[63355]: value = "task-1349584" [ 854.041165] env[63355]: _type = "Task" [ 854.041165] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.048931] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349584, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.461989] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Releasing lock "refresh_cache-7d79c9cf-66a9-4117-b090-0ab0676c3114" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.462261] env[63355]: DEBUG nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 854.462450] env[63355]: DEBUG nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.462622] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.478393] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.505211] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 854.505387] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Cleaning up deleted instances with incomplete migration {{(pid=63355) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 854.549918] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349584, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447213} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.552229] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.552449] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.552858] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-faf02f0f-5422-4245-8c49-555fad0b1bd4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.559340] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for the task: (returnval){ [ 854.559340] env[63355]: value = "task-1349585" [ 854.559340] env[63355]: _type = "Task" [ 854.559340] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.568334] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349585, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.646801] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000f540c-e9f2-49f2-a290-cd256ee9a598 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.654797] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2a708f-346c-4547-a4c6-233ffde37ef3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.414078] env[63355]: DEBUG nova.network.neutron [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.415877] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 855.419955] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ec58b3-7944-4b20-bc1d-772415d6fc84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.428944] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349585, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061348} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.431139] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.432442] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d285879e-fee4-45ba-9526-01df99094ad8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.436151] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80de415-ebfb-4eb6-98c0-27b13e42cde4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.456928] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.464779] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59edada8-9d90-4344-b835-f8e109698503 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.478482] env[63355]: DEBUG nova.compute.provider_tree [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.485485] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for the task: (returnval){ [ 855.485485] env[63355]: value = "task-1349586" [ 855.485485] env[63355]: _type = "Task" [ 855.485485] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.493876] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349586, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.917261] env[63355]: INFO nova.compute.manager [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] [instance: 7d79c9cf-66a9-4117-b090-0ab0676c3114] Took 1.45 seconds to deallocate network for instance. [ 855.981267] env[63355]: DEBUG nova.scheduler.client.report [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.995404] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349586, 'name': ReconfigVM_Task, 'duration_secs': 0.402976} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.996208] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70/9989c129-07d3-4af9-9a86-1e2746d6ac70.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.996774] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de4324a5-13a9-42f6-9d4f-99f5b8ac4773 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.003052] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for the task: (returnval){ [ 856.003052] env[63355]: value = "task-1349587" [ 856.003052] env[63355]: _type = "Task" [ 856.003052] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.011133] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349587, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.486015] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.144s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.486621] env[63355]: DEBUG nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.489250] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.957s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.513627] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349587, 'name': Rename_Task, 'duration_secs': 0.132421} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.513879] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.514124] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-062f861b-01bc-45d1-9717-4d56c8e6dade {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.521052] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Waiting for the task: (returnval){ [ 856.521052] env[63355]: value = "task-1349588" [ 856.521052] env[63355]: _type = "Task" [ 856.521052] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.528584] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349588, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.945671] env[63355]: INFO nova.scheduler.client.report [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Deleted allocations for instance 7d79c9cf-66a9-4117-b090-0ab0676c3114 [ 856.995988] env[63355]: DEBUG nova.compute.utils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.997285] env[63355]: DEBUG nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 856.997450] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 857.033396] env[63355]: DEBUG oslo_vmware.api [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Task: {'id': task-1349588, 'name': PowerOnVM_Task, 'duration_secs': 0.400324} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.035825] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.036053] env[63355]: DEBUG nova.compute.manager [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.037157] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7c3bef-e9f0-49c6-b306-4d284b474044 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.041801] env[63355]: DEBUG nova.policy [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0f1d54b2d9c454688e7f56a40119657', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad0fba7800a641da9114e93e18d7977d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.296466] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8810397-4604-417e-9d15-a1c6cc35248d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.304446] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bd59e3-f3b7-49de-a09f-7e413ffa044a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.333286] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Successfully created port: a83881a1-bb59-41b5-b08c-7999d67d97af {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.335212] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d292c5d3-740a-4fd4-bc18-28b9c2857c80 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.343029] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94769ffe-66a0-426e-b8a7-16c195d66b1b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.356740] env[63355]: DEBUG nova.compute.provider_tree [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.456717] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c6d15249-3f8b-4e4f-948e-6fbcae9ebc76 tempest-ServerRescueNegativeTestJSON-1344519775 tempest-ServerRescueNegativeTestJSON-1344519775-project-member] Lock "7d79c9cf-66a9-4117-b090-0ab0676c3114" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 187.231s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.500254] env[63355]: DEBUG nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 857.555420] env[63355]: DEBUG oslo_concurrency.lockutils [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.640276] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Successfully created port: 0da31491-1df8-4db7-abe5-3df49a5e3d23 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.862251] env[63355]: DEBUG nova.scheduler.client.report [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.960342] env[63355]: DEBUG nova.compute.manager [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.042107] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "9989c129-07d3-4af9-9a86-1e2746d6ac70" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.042369] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "9989c129-07d3-4af9-9a86-1e2746d6ac70" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.042577] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "9989c129-07d3-4af9-9a86-1e2746d6ac70-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.042754] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "9989c129-07d3-4af9-9a86-1e2746d6ac70-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.042920] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "9989c129-07d3-4af9-9a86-1e2746d6ac70-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.046345] env[63355]: INFO nova.compute.manager [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Terminating instance [ 858.048474] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "refresh_cache-9989c129-07d3-4af9-9a86-1e2746d6ac70" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.048632] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquired lock "refresh_cache-9989c129-07d3-4af9-9a86-1e2746d6ac70" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.048793] env[63355]: DEBUG nova.network.neutron [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.231837] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Successfully created port: d4ac2f82-ac73-464f-ae2f-e7d648b6ad98 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.366609] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.877s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.367268] env[63355]: ERROR nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 098f7509-6158-4f8f-bc97-63435fa49c6f, please check neutron logs for more information. [ 858.367268] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Traceback (most recent call last): [ 858.367268] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 858.367268] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] self.driver.spawn(context, instance, image_meta, [ 858.367268] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 858.367268] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 858.367268] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 858.367268] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] vm_ref = self.build_virtual_machine(instance, [ 858.367268] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 858.367268] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] vif_infos = vmwarevif.get_vif_info(self._session, [ 858.367268] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] for vif in network_info: [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] return self._sync_wrapper(fn, *args, **kwargs) [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] self.wait() [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] self[:] = self._gt.wait() [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] return self._exit_event.wait() [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] current.throw(*self._exc) [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 858.367776] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] result = function(*args, **kwargs) [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] return func(*args, **kwargs) [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] raise e [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] nwinfo = self.network_api.allocate_for_instance( [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] created_port_ids = self._update_ports_for_instance( [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] with excutils.save_and_reraise_exception(): [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] self.force_reraise() [ 858.368283] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 858.368843] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] raise self.value [ 858.368843] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 858.368843] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] updated_port = self._update_port( [ 858.368843] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 858.368843] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] _ensure_no_port_binding_failure(port) [ 858.368843] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 858.368843] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] raise exception.PortBindingFailed(port_id=port['id']) [ 858.368843] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] nova.exception.PortBindingFailed: Binding failed for port 098f7509-6158-4f8f-bc97-63435fa49c6f, please check neutron logs for more information. [ 858.368843] env[63355]: ERROR nova.compute.manager [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] [ 858.368843] env[63355]: DEBUG nova.compute.utils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Binding failed for port 098f7509-6158-4f8f-bc97-63435fa49c6f, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 858.369274] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.910s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.370677] env[63355]: INFO nova.compute.claims [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.374747] env[63355]: DEBUG nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Build of instance 4ded6053-9afc-444e-bdfc-244265ed19a1 was re-scheduled: Binding failed for port 098f7509-6158-4f8f-bc97-63435fa49c6f, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 858.374747] env[63355]: DEBUG nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 858.374747] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquiring lock "refresh_cache-4ded6053-9afc-444e-bdfc-244265ed19a1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.374747] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquired lock "refresh_cache-4ded6053-9afc-444e-bdfc-244265ed19a1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.375520] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.482042] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.510681] env[63355]: DEBUG nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.550536] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.551432] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.551432] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.551432] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.551432] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.551432] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.551712] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.551712] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.551869] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.552033] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.552200] env[63355]: DEBUG nova.virt.hardware [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.554863] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cfc4c04-5db2-40d9-bd89-236563a12032 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.563834] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ea68fb-8f09-473d-bfab-b3768452e5e5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.579387] env[63355]: DEBUG nova.network.neutron [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.642195] env[63355]: DEBUG nova.network.neutron [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.905578] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 859.021341] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.145268] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Releasing lock "refresh_cache-9989c129-07d3-4af9-9a86-1e2746d6ac70" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.145736] env[63355]: DEBUG nova.compute.manager [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 859.145931] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.146819] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b234df8a-cbff-44fd-903b-5cc6eaeb5010 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.155208] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.155447] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f761443-bb56-4785-924d-b4564ee3dfe5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.161046] env[63355]: DEBUG oslo_vmware.api [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 859.161046] env[63355]: value = "task-1349589" [ 859.161046] env[63355]: _type = "Task" [ 859.161046] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.169232] env[63355]: DEBUG oslo_vmware.api [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349589, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.440448] env[63355]: DEBUG nova.compute.manager [req-9a0e12ad-c7d7-49c8-a124-97b9346a5405 req-4739878e-ec02-4910-aca4-821f9a198858 service nova] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Received event network-changed-a83881a1-bb59-41b5-b08c-7999d67d97af {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.440654] env[63355]: DEBUG nova.compute.manager [req-9a0e12ad-c7d7-49c8-a124-97b9346a5405 req-4739878e-ec02-4910-aca4-821f9a198858 service nova] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Refreshing instance network info cache due to event network-changed-a83881a1-bb59-41b5-b08c-7999d67d97af. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 859.440862] env[63355]: DEBUG oslo_concurrency.lockutils [req-9a0e12ad-c7d7-49c8-a124-97b9346a5405 req-4739878e-ec02-4910-aca4-821f9a198858 service nova] Acquiring lock "refresh_cache-e49ba26c-486c-43aa-9595-231e7f31a522" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.441057] env[63355]: DEBUG oslo_concurrency.lockutils [req-9a0e12ad-c7d7-49c8-a124-97b9346a5405 req-4739878e-ec02-4910-aca4-821f9a198858 service nova] Acquired lock "refresh_cache-e49ba26c-486c-43aa-9595-231e7f31a522" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.441181] env[63355]: DEBUG nova.network.neutron [req-9a0e12ad-c7d7-49c8-a124-97b9346a5405 req-4739878e-ec02-4910-aca4-821f9a198858 service nova] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Refreshing network info cache for port a83881a1-bb59-41b5-b08c-7999d67d97af {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 859.523656] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Releasing lock "refresh_cache-4ded6053-9afc-444e-bdfc-244265ed19a1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.523930] env[63355]: DEBUG nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 859.524051] env[63355]: DEBUG nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 859.524215] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 859.542512] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 859.561535] env[63355]: ERROR nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a83881a1-bb59-41b5-b08c-7999d67d97af, please check neutron logs for more information. [ 859.561535] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 859.561535] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 859.561535] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 859.561535] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 859.561535] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 859.561535] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 859.561535] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 859.561535] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 859.561535] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 859.561535] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 859.561535] env[63355]: ERROR nova.compute.manager raise self.value [ 859.561535] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 859.561535] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 859.561535] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 859.561535] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 859.562007] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 859.562007] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 859.562007] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a83881a1-bb59-41b5-b08c-7999d67d97af, please check neutron logs for more information. [ 859.562007] env[63355]: ERROR nova.compute.manager [ 859.562007] env[63355]: Traceback (most recent call last): [ 859.562007] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 859.562007] env[63355]: listener.cb(fileno) [ 859.562007] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 859.562007] env[63355]: result = function(*args, **kwargs) [ 859.562007] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 859.562007] env[63355]: return func(*args, **kwargs) [ 859.562007] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 859.562007] env[63355]: raise e [ 859.562007] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 859.562007] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 859.562007] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 859.562007] env[63355]: created_port_ids = self._update_ports_for_instance( [ 859.562007] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 859.562007] env[63355]: with excutils.save_and_reraise_exception(): [ 859.562007] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 859.562007] env[63355]: self.force_reraise() [ 859.562007] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 859.562007] env[63355]: raise self.value [ 859.562007] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 859.562007] env[63355]: updated_port = self._update_port( [ 859.562007] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 859.562007] env[63355]: _ensure_no_port_binding_failure(port) [ 859.562007] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 859.562007] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 859.562760] env[63355]: nova.exception.PortBindingFailed: Binding failed for port a83881a1-bb59-41b5-b08c-7999d67d97af, please check neutron logs for more information. [ 859.562760] env[63355]: Removing descriptor: 17 [ 859.562760] env[63355]: ERROR nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a83881a1-bb59-41b5-b08c-7999d67d97af, please check neutron logs for more information. [ 859.562760] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Traceback (most recent call last): [ 859.562760] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 859.562760] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] yield resources [ 859.562760] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 859.562760] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] self.driver.spawn(context, instance, image_meta, [ 859.562760] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 859.562760] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] self._vmops.spawn(context, instance, image_meta, injected_files, [ 859.562760] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 859.562760] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] vm_ref = self.build_virtual_machine(instance, [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] vif_infos = vmwarevif.get_vif_info(self._session, [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] for vif in network_info: [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] return self._sync_wrapper(fn, *args, **kwargs) [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] self.wait() [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] self[:] = self._gt.wait() [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] return self._exit_event.wait() [ 859.563108] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] result = hub.switch() [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] return self.greenlet.switch() [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] result = function(*args, **kwargs) [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] return func(*args, **kwargs) [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] raise e [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] nwinfo = self.network_api.allocate_for_instance( [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 859.563457] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] created_port_ids = self._update_ports_for_instance( [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] with excutils.save_and_reraise_exception(): [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] self.force_reraise() [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] raise self.value [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] updated_port = self._update_port( [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] _ensure_no_port_binding_failure(port) [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 859.563737] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] raise exception.PortBindingFailed(port_id=port['id']) [ 859.563997] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] nova.exception.PortBindingFailed: Binding failed for port a83881a1-bb59-41b5-b08c-7999d67d97af, please check neutron logs for more information. [ 859.563997] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] [ 859.563997] env[63355]: INFO nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Terminating instance [ 859.564692] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "refresh_cache-e49ba26c-486c-43aa-9595-231e7f31a522" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.671335] env[63355]: DEBUG oslo_vmware.api [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349589, 'name': PowerOffVM_Task, 'duration_secs': 0.183173} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.673755] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 859.673930] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 859.675048] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c57bbd5-b130-4a7a-a3a5-41dd4b3d7ff3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.700547] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 859.700765] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 859.700940] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Deleting the datastore file [datastore2] 9989c129-07d3-4af9-9a86-1e2746d6ac70 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 859.701207] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-201a06d1-0c07-454e-8445-992d1ba6eb85 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.709082] env[63355]: DEBUG oslo_vmware.api [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for the task: (returnval){ [ 859.709082] env[63355]: value = "task-1349591" [ 859.709082] env[63355]: _type = "Task" [ 859.709082] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.716758] env[63355]: DEBUG oslo_vmware.api [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349591, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.777058] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9831ec31-c1e6-4de2-8b0f-f3ae0f944bc6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.784771] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e06b849-c25c-46d6-8cad-bdf6602c61bc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.814955] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd56e97f-7eb6-4482-9947-e8206d83ff52 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.822347] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cb32f9-63d6-41d7-9922-8061d62287ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.835116] env[63355]: DEBUG nova.compute.provider_tree [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.998802] env[63355]: DEBUG nova.network.neutron [req-9a0e12ad-c7d7-49c8-a124-97b9346a5405 req-4739878e-ec02-4910-aca4-821f9a198858 service nova] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.048544] env[63355]: DEBUG nova.network.neutron [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.091997] env[63355]: DEBUG nova.network.neutron [req-9a0e12ad-c7d7-49c8-a124-97b9346a5405 req-4739878e-ec02-4910-aca4-821f9a198858 service nova] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.224354] env[63355]: DEBUG oslo_vmware.api [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Task: {'id': task-1349591, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097761} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.224814] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.225098] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.225368] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.225607] env[63355]: INFO nova.compute.manager [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Took 1.08 seconds to destroy the instance on the hypervisor. [ 860.225904] env[63355]: DEBUG oslo.service.loopingcall [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.226163] env[63355]: DEBUG nova.compute.manager [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 860.226316] env[63355]: DEBUG nova.network.neutron [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 860.241949] env[63355]: DEBUG nova.network.neutron [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.337811] env[63355]: DEBUG nova.scheduler.client.report [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.550945] env[63355]: INFO nova.compute.manager [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: 4ded6053-9afc-444e-bdfc-244265ed19a1] Took 1.03 seconds to deallocate network for instance. [ 860.596462] env[63355]: DEBUG oslo_concurrency.lockutils [req-9a0e12ad-c7d7-49c8-a124-97b9346a5405 req-4739878e-ec02-4910-aca4-821f9a198858 service nova] Releasing lock "refresh_cache-e49ba26c-486c-43aa-9595-231e7f31a522" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.596865] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquired lock "refresh_cache-e49ba26c-486c-43aa-9595-231e7f31a522" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.597069] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.744585] env[63355]: DEBUG nova.network.neutron [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.843163] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.843677] env[63355]: DEBUG nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 860.846128] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.533s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.847572] env[63355]: INFO nova.compute.claims [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.134095] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.212129] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.247110] env[63355]: INFO nova.compute.manager [-] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Took 1.02 seconds to deallocate network for instance. [ 861.353284] env[63355]: DEBUG nova.compute.utils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 861.354661] env[63355]: DEBUG nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 861.354869] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 861.400384] env[63355]: DEBUG nova.policy [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'beb41983e69a44b2a108f552da5ef640', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfff1b9903264e5586119ebd3a3602de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 861.468995] env[63355]: DEBUG nova.compute.manager [req-04cad3c2-c1ba-4fc4-b6cf-4e482a7925c8 req-431b012f-edea-4729-b0f2-d1bca4238274 service nova] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Received event network-vif-deleted-a83881a1-bb59-41b5-b08c-7999d67d97af {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.590798] env[63355]: INFO nova.scheduler.client.report [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Deleted allocations for instance 4ded6053-9afc-444e-bdfc-244265ed19a1 [ 861.651355] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Successfully created port: 5a14c6da-f788-43f9-826a-4ce8b677fe35 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 861.715586] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Releasing lock "refresh_cache-e49ba26c-486c-43aa-9595-231e7f31a522" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.715817] env[63355]: DEBUG nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 861.716105] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 861.716657] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4da5547-935d-4f61-a1c6-d60e0f11229b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.725871] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c59b7b2-4937-48bd-b9e2-e6183ce07b21 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.747166] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e49ba26c-486c-43aa-9595-231e7f31a522 could not be found. [ 861.747386] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 861.747613] env[63355]: INFO nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Took 0.03 seconds to destroy the instance on the hypervisor. [ 861.747911] env[63355]: DEBUG oslo.service.loopingcall [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.748135] env[63355]: DEBUG nova.compute.manager [-] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 861.748229] env[63355]: DEBUG nova.network.neutron [-] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 861.754746] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.779186] env[63355]: DEBUG nova.network.neutron [-] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.859635] env[63355]: DEBUG nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 862.099804] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0a8428c-5a39-4d26-be2d-a416bda593f0 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "4ded6053-9afc-444e-bdfc-244265ed19a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 183.068s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.149993] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529c255b-8e45-4cfe-b7d6-c3bb50ff4a13 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.157870] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3da1d1-91bb-40c3-9b43-b1f9b6dbc91e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.188846] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d15a6e-5660-4b0f-b62e-59cb07177bb8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.195996] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93ecde8-91e3-4d1c-8dd8-a76bb41c0f7b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.209032] env[63355]: DEBUG nova.compute.provider_tree [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.538662] env[63355]: ERROR nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5a14c6da-f788-43f9-826a-4ce8b677fe35, please check neutron logs for more information. [ 862.538662] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 862.538662] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 862.538662] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 862.538662] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 862.538662] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 862.538662] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 862.538662] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 862.538662] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 862.538662] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 862.538662] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 862.538662] env[63355]: ERROR nova.compute.manager raise self.value [ 862.538662] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 862.538662] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 862.538662] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 862.538662] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 862.539323] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 862.539323] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 862.539323] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5a14c6da-f788-43f9-826a-4ce8b677fe35, please check neutron logs for more information. [ 862.539323] env[63355]: ERROR nova.compute.manager [ 862.539323] env[63355]: Traceback (most recent call last): [ 862.539323] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 862.539323] env[63355]: listener.cb(fileno) [ 862.539323] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 862.539323] env[63355]: result = function(*args, **kwargs) [ 862.539323] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 862.539323] env[63355]: return func(*args, **kwargs) [ 862.539323] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 862.539323] env[63355]: raise e [ 862.539323] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 862.539323] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 862.539323] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 862.539323] env[63355]: created_port_ids = self._update_ports_for_instance( [ 862.539323] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 862.539323] env[63355]: with excutils.save_and_reraise_exception(): [ 862.539323] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 862.539323] env[63355]: self.force_reraise() [ 862.539323] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 862.539323] env[63355]: raise self.value [ 862.539323] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 862.539323] env[63355]: updated_port = self._update_port( [ 862.539323] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 862.539323] env[63355]: _ensure_no_port_binding_failure(port) [ 862.539323] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 862.539323] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 862.540402] env[63355]: nova.exception.PortBindingFailed: Binding failed for port 5a14c6da-f788-43f9-826a-4ce8b677fe35, please check neutron logs for more information. [ 862.540402] env[63355]: Removing descriptor: 17 [ 862.602745] env[63355]: DEBUG nova.compute.manager [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 862.712388] env[63355]: DEBUG nova.scheduler.client.report [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.869832] env[63355]: DEBUG nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 862.893290] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.893547] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.893703] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.893885] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.894045] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.894215] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.894420] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.894576] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.894737] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.894898] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.895081] env[63355]: DEBUG nova.virt.hardware [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.895974] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc1187b-db1d-403d-98ff-3bad23dbc48e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.904112] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0aa780-f35d-4dec-a8e2-5e0a7139458e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.918509] env[63355]: ERROR nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5a14c6da-f788-43f9-826a-4ce8b677fe35, please check neutron logs for more information. [ 862.918509] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Traceback (most recent call last): [ 862.918509] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 862.918509] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] yield resources [ 862.918509] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 862.918509] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] self.driver.spawn(context, instance, image_meta, [ 862.918509] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 862.918509] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 862.918509] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 862.918509] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] vm_ref = self.build_virtual_machine(instance, [ 862.918509] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] vif_infos = vmwarevif.get_vif_info(self._session, [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] for vif in network_info: [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] return self._sync_wrapper(fn, *args, **kwargs) [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] self.wait() [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] self[:] = self._gt.wait() [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] return self._exit_event.wait() [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 862.918825] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] current.throw(*self._exc) [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] result = function(*args, **kwargs) [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] return func(*args, **kwargs) [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] raise e [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] nwinfo = self.network_api.allocate_for_instance( [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] created_port_ids = self._update_ports_for_instance( [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] with excutils.save_and_reraise_exception(): [ 862.919183] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] self.force_reraise() [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] raise self.value [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] updated_port = self._update_port( [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] _ensure_no_port_binding_failure(port) [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] raise exception.PortBindingFailed(port_id=port['id']) [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] nova.exception.PortBindingFailed: Binding failed for port 5a14c6da-f788-43f9-826a-4ce8b677fe35, please check neutron logs for more information. [ 862.919523] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] [ 862.919523] env[63355]: INFO nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Terminating instance [ 862.921019] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-7dd28cb7-d249-4b8e-beb3-c959f15e912f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.921019] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-7dd28cb7-d249-4b8e-beb3-c959f15e912f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.921129] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 862.976345] env[63355]: DEBUG nova.network.neutron [-] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.126824] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.217165] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.217924] env[63355]: DEBUG nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 863.220676] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.186s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.439951] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 863.479027] env[63355]: INFO nova.compute.manager [-] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Took 1.73 seconds to deallocate network for instance. [ 863.480782] env[63355]: DEBUG nova.compute.claims [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 863.480962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.495575] env[63355]: DEBUG nova.compute.manager [req-ad37f3ba-3310-4e1c-87b8-e2b32be00e44 req-29a7c9a6-465e-45de-a042-5a28305d9ceb service nova] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Received event network-changed-5a14c6da-f788-43f9-826a-4ce8b677fe35 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.495782] env[63355]: DEBUG nova.compute.manager [req-ad37f3ba-3310-4e1c-87b8-e2b32be00e44 req-29a7c9a6-465e-45de-a042-5a28305d9ceb service nova] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Refreshing instance network info cache due to event network-changed-5a14c6da-f788-43f9-826a-4ce8b677fe35. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 863.495970] env[63355]: DEBUG oslo_concurrency.lockutils [req-ad37f3ba-3310-4e1c-87b8-e2b32be00e44 req-29a7c9a6-465e-45de-a042-5a28305d9ceb service nova] Acquiring lock "refresh_cache-7dd28cb7-d249-4b8e-beb3-c959f15e912f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.526310] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.732281] env[63355]: DEBUG nova.compute.utils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 863.733828] env[63355]: DEBUG nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 863.733996] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 863.778058] env[63355]: DEBUG nova.policy [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bf70ebeeb3a4327ada51d18f5651476', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce0682a99ac94aeea463c961b84e6b58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.028649] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-7dd28cb7-d249-4b8e-beb3-c959f15e912f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.029246] env[63355]: DEBUG nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 864.029474] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.029798] env[63355]: DEBUG oslo_concurrency.lockutils [req-ad37f3ba-3310-4e1c-87b8-e2b32be00e44 req-29a7c9a6-465e-45de-a042-5a28305d9ceb service nova] Acquired lock "refresh_cache-7dd28cb7-d249-4b8e-beb3-c959f15e912f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.029982] env[63355]: DEBUG nova.network.neutron [req-ad37f3ba-3310-4e1c-87b8-e2b32be00e44 req-29a7c9a6-465e-45de-a042-5a28305d9ceb service nova] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Refreshing network info cache for port 5a14c6da-f788-43f9-826a-4ce8b677fe35 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 864.031389] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-307cc6d7-9b56-442b-9e63-d8e26ae6cb20 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.042194] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d4ce7a-e014-4c7a-a65d-70add6a18321 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.060953] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960e3c48-1d40-49f3-9e10-2450a9ff8b75 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.079980] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Successfully created port: cb86c9bd-d267-4b95-a0ff-f75747ef143f {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 864.084262] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e00e8a-423b-4109-8d6d-107b3b22d3af {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.125912] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6f18f9-4c7f-4b69-934d-e399d800805f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.128766] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7dd28cb7-d249-4b8e-beb3-c959f15e912f could not be found. [ 864.128983] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.129179] env[63355]: INFO nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Took 0.10 seconds to destroy the instance on the hypervisor. [ 864.129517] env[63355]: DEBUG oslo.service.loopingcall [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.129739] env[63355]: DEBUG nova.compute.manager [-] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.129834] env[63355]: DEBUG nova.network.neutron [-] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 864.135767] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23325f91-5aa8-4df7-a0c9-62d1b87e0f1c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.150561] env[63355]: DEBUG nova.compute.provider_tree [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.154326] env[63355]: DEBUG nova.network.neutron [-] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.236788] env[63355]: DEBUG nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 864.561784] env[63355]: DEBUG nova.network.neutron [req-ad37f3ba-3310-4e1c-87b8-e2b32be00e44 req-29a7c9a6-465e-45de-a042-5a28305d9ceb service nova] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.653903] env[63355]: DEBUG nova.network.neutron [req-ad37f3ba-3310-4e1c-87b8-e2b32be00e44 req-29a7c9a6-465e-45de-a042-5a28305d9ceb service nova] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.655717] env[63355]: DEBUG nova.scheduler.client.report [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.659851] env[63355]: DEBUG nova.network.neutron [-] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.032791] env[63355]: ERROR nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cb86c9bd-d267-4b95-a0ff-f75747ef143f, please check neutron logs for more information. [ 865.032791] env[63355]: ERROR nova.compute.manager Traceback (most recent call last): [ 865.032791] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 865.032791] env[63355]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 865.032791] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 865.032791] env[63355]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 865.032791] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 865.032791] env[63355]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 865.032791] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 865.032791] env[63355]: ERROR nova.compute.manager self.force_reraise() [ 865.032791] env[63355]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 865.032791] env[63355]: ERROR nova.compute.manager raise self.value [ 865.032791] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 865.032791] env[63355]: ERROR nova.compute.manager updated_port = self._update_port( [ 865.032791] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 865.032791] env[63355]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 865.033506] env[63355]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 865.033506] env[63355]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 865.033506] env[63355]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cb86c9bd-d267-4b95-a0ff-f75747ef143f, please check neutron logs for more information. [ 865.033506] env[63355]: ERROR nova.compute.manager [ 865.033506] env[63355]: Traceback (most recent call last): [ 865.033506] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 865.033506] env[63355]: listener.cb(fileno) [ 865.033506] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 865.033506] env[63355]: result = function(*args, **kwargs) [ 865.033506] env[63355]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 865.033506] env[63355]: return func(*args, **kwargs) [ 865.033506] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 865.033506] env[63355]: raise e [ 865.033506] env[63355]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 865.033506] env[63355]: nwinfo = self.network_api.allocate_for_instance( [ 865.033506] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 865.033506] env[63355]: created_port_ids = self._update_ports_for_instance( [ 865.033506] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 865.033506] env[63355]: with excutils.save_and_reraise_exception(): [ 865.033506] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 865.033506] env[63355]: self.force_reraise() [ 865.033506] env[63355]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 865.033506] env[63355]: raise self.value [ 865.033506] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 865.033506] env[63355]: updated_port = self._update_port( [ 865.033506] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 865.033506] env[63355]: _ensure_no_port_binding_failure(port) [ 865.033506] env[63355]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 865.033506] env[63355]: raise exception.PortBindingFailed(port_id=port['id']) [ 865.034258] env[63355]: nova.exception.PortBindingFailed: Binding failed for port cb86c9bd-d267-4b95-a0ff-f75747ef143f, please check neutron logs for more information. [ 865.034258] env[63355]: Removing descriptor: 17 [ 865.160598] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.940s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.161235] env[63355]: ERROR nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 33d8e84f-04c9-4513-961a-9deff4a565dd, please check neutron logs for more information. [ 865.161235] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Traceback (most recent call last): [ 865.161235] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 865.161235] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] self.driver.spawn(context, instance, image_meta, [ 865.161235] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 865.161235] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 865.161235] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 865.161235] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] vm_ref = self.build_virtual_machine(instance, [ 865.161235] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 865.161235] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] vif_infos = vmwarevif.get_vif_info(self._session, [ 865.161235] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] for vif in network_info: [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] return self._sync_wrapper(fn, *args, **kwargs) [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] self.wait() [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] self[:] = self._gt.wait() [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] return self._exit_event.wait() [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] current.throw(*self._exc) [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 865.161701] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] result = function(*args, **kwargs) [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] return func(*args, **kwargs) [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] raise e [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] nwinfo = self.network_api.allocate_for_instance( [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] created_port_ids = self._update_ports_for_instance( [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] with excutils.save_and_reraise_exception(): [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] self.force_reraise() [ 865.162089] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 865.162380] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] raise self.value [ 865.162380] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 865.162380] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] updated_port = self._update_port( [ 865.162380] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 865.162380] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] _ensure_no_port_binding_failure(port) [ 865.162380] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 865.162380] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] raise exception.PortBindingFailed(port_id=port['id']) [ 865.162380] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] nova.exception.PortBindingFailed: Binding failed for port 33d8e84f-04c9-4513-961a-9deff4a565dd, please check neutron logs for more information. [ 865.162380] env[63355]: ERROR nova.compute.manager [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] [ 865.162380] env[63355]: DEBUG nova.compute.utils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Binding failed for port 33d8e84f-04c9-4513-961a-9deff4a565dd, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 865.163321] env[63355]: DEBUG oslo_concurrency.lockutils [req-ad37f3ba-3310-4e1c-87b8-e2b32be00e44 req-29a7c9a6-465e-45de-a042-5a28305d9ceb service nova] Releasing lock "refresh_cache-7dd28cb7-d249-4b8e-beb3-c959f15e912f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.163539] env[63355]: DEBUG nova.compute.manager [req-ad37f3ba-3310-4e1c-87b8-e2b32be00e44 req-29a7c9a6-465e-45de-a042-5a28305d9ceb service nova] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Received event network-vif-deleted-5a14c6da-f788-43f9-826a-4ce8b677fe35 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.164098] env[63355]: DEBUG nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Build of instance dd80c3be-ec92-4a67-a373-3e2276752feb was re-scheduled: Binding failed for port 33d8e84f-04c9-4513-961a-9deff4a565dd, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 865.164497] env[63355]: DEBUG nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 865.164764] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquiring lock "refresh_cache-dd80c3be-ec92-4a67-a373-3e2276752feb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.164928] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Acquired lock "refresh_cache-dd80c3be-ec92-4a67-a373-3e2276752feb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.165110] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 865.166086] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.008s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.169351] env[63355]: INFO nova.compute.claims [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.170892] env[63355]: INFO nova.compute.manager [-] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Took 1.04 seconds to deallocate network for instance. [ 865.172593] env[63355]: DEBUG nova.compute.claims [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 865.172721] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.245614] env[63355]: DEBUG nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 865.270413] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 865.270896] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 865.270896] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.271025] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 865.271154] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.271314] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 865.271502] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 865.272059] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 865.272059] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 865.272059] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 865.272201] env[63355]: DEBUG nova.virt.hardware [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.272968] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d76a66b-62bd-47c0-82fa-76f791536eca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.281103] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c26de5-4a91-4a90-b08f-70bfee1a9f07 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.301048] env[63355]: ERROR nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cb86c9bd-d267-4b95-a0ff-f75747ef143f, please check neutron logs for more information. [ 865.301048] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Traceback (most recent call last): [ 865.301048] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 865.301048] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] yield resources [ 865.301048] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 865.301048] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] self.driver.spawn(context, instance, image_meta, [ 865.301048] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 865.301048] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 865.301048] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 865.301048] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] vm_ref = self.build_virtual_machine(instance, [ 865.301048] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] vif_infos = vmwarevif.get_vif_info(self._session, [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] for vif in network_info: [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] return self._sync_wrapper(fn, *args, **kwargs) [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] self.wait() [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] self[:] = self._gt.wait() [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] return self._exit_event.wait() [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 865.301511] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] current.throw(*self._exc) [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] result = function(*args, **kwargs) [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] return func(*args, **kwargs) [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] raise e [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] nwinfo = self.network_api.allocate_for_instance( [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] created_port_ids = self._update_ports_for_instance( [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] with excutils.save_and_reraise_exception(): [ 865.301865] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] self.force_reraise() [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] raise self.value [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] updated_port = self._update_port( [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] _ensure_no_port_binding_failure(port) [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] raise exception.PortBindingFailed(port_id=port['id']) [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] nova.exception.PortBindingFailed: Binding failed for port cb86c9bd-d267-4b95-a0ff-f75747ef143f, please check neutron logs for more information. [ 865.302202] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] [ 865.302202] env[63355]: INFO nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Terminating instance [ 865.303468] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-947a33d3-88c1-45ae-805c-eda2bc7626f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.303629] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-947a33d3-88c1-45ae-805c-eda2bc7626f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.303791] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 865.552708] env[63355]: DEBUG nova.compute.manager [req-876a6b6c-71d4-475f-999b-9d4ec21a6992 req-cb35e588-6888-4017-b90a-91bfd38d0caf service nova] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Received event network-changed-cb86c9bd-d267-4b95-a0ff-f75747ef143f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.552913] env[63355]: DEBUG nova.compute.manager [req-876a6b6c-71d4-475f-999b-9d4ec21a6992 req-cb35e588-6888-4017-b90a-91bfd38d0caf service nova] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Refreshing instance network info cache due to event network-changed-cb86c9bd-d267-4b95-a0ff-f75747ef143f. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.553119] env[63355]: DEBUG oslo_concurrency.lockutils [req-876a6b6c-71d4-475f-999b-9d4ec21a6992 req-cb35e588-6888-4017-b90a-91bfd38d0caf service nova] Acquiring lock "refresh_cache-947a33d3-88c1-45ae-805c-eda2bc7626f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.701345] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 865.794312] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.824175] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 865.903903] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.296892] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Releasing lock "refresh_cache-dd80c3be-ec92-4a67-a373-3e2276752feb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.297176] env[63355]: DEBUG nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 866.297311] env[63355]: DEBUG nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 866.297485] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 866.312656] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.406058] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-947a33d3-88c1-45ae-805c-eda2bc7626f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.406633] env[63355]: DEBUG nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 866.406850] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.407192] env[63355]: DEBUG oslo_concurrency.lockutils [req-876a6b6c-71d4-475f-999b-9d4ec21a6992 req-cb35e588-6888-4017-b90a-91bfd38d0caf service nova] Acquired lock "refresh_cache-947a33d3-88c1-45ae-805c-eda2bc7626f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.407373] env[63355]: DEBUG nova.network.neutron [req-876a6b6c-71d4-475f-999b-9d4ec21a6992 req-cb35e588-6888-4017-b90a-91bfd38d0caf service nova] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Refreshing network info cache for port cb86c9bd-d267-4b95-a0ff-f75747ef143f {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 866.408402] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-943eb2fe-38d4-432a-a7c5-b39ea2a2a7db {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.413682] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d6bfe7-6d49-4dfd-930f-9bb2413fe37b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.420112] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb216f8-5825-4649-9260-5c069997cf75 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.432650] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19bf79a9-af53-4879-a521-4fb4127020a3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.466963] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf2f7c22-4e90-49a1-ab9f-0fb9e31cc289 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.469693] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 947a33d3-88c1-45ae-805c-eda2bc7626f1 could not be found. [ 866.469905] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 866.470106] env[63355]: INFO nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Took 0.06 seconds to destroy the instance on the hypervisor. [ 866.470346] env[63355]: DEBUG oslo.service.loopingcall [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.470560] env[63355]: DEBUG nova.compute.manager [-] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 866.470653] env[63355]: DEBUG nova.network.neutron [-] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 866.476934] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d8cc25-93e7-46bb-bf92-22756de73bd2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.491727] env[63355]: DEBUG nova.compute.provider_tree [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.494077] env[63355]: DEBUG nova.network.neutron [-] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.816108] env[63355]: DEBUG nova.network.neutron [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.926712] env[63355]: DEBUG nova.network.neutron [req-876a6b6c-71d4-475f-999b-9d4ec21a6992 req-cb35e588-6888-4017-b90a-91bfd38d0caf service nova] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.996227] env[63355]: DEBUG nova.scheduler.client.report [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.999537] env[63355]: DEBUG nova.network.neutron [-] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.001099] env[63355]: DEBUG nova.network.neutron [req-876a6b6c-71d4-475f-999b-9d4ec21a6992 req-cb35e588-6888-4017-b90a-91bfd38d0caf service nova] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.318342] env[63355]: INFO nova.compute.manager [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] [instance: dd80c3be-ec92-4a67-a373-3e2276752feb] Took 1.02 seconds to deallocate network for instance. [ 867.504136] env[63355]: DEBUG oslo_concurrency.lockutils [req-876a6b6c-71d4-475f-999b-9d4ec21a6992 req-cb35e588-6888-4017-b90a-91bfd38d0caf service nova] Releasing lock "refresh_cache-947a33d3-88c1-45ae-805c-eda2bc7626f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.504136] env[63355]: DEBUG nova.compute.manager [req-876a6b6c-71d4-475f-999b-9d4ec21a6992 req-cb35e588-6888-4017-b90a-91bfd38d0caf service nova] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Received event network-vif-deleted-cb86c9bd-d267-4b95-a0ff-f75747ef143f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.504630] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.505226] env[63355]: DEBUG nova.compute.manager [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 867.507728] env[63355]: INFO nova.compute.manager [-] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Took 1.04 seconds to deallocate network for instance. [ 867.508061] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.040s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.509513] env[63355]: INFO nova.compute.claims [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.515704] env[63355]: DEBUG nova.compute.claims [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Aborting claim: {{(pid=63355) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 867.515930] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.015851] env[63355]: DEBUG nova.compute.utils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.019490] env[63355]: DEBUG nova.compute.manager [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.019744] env[63355]: DEBUG nova.network.neutron [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 868.057880] env[63355]: DEBUG nova.policy [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fbc2326b0af476e91c000e1b123eabe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a26599cf04b24e17aac6f682b641c703', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 868.305088] env[63355]: DEBUG nova.network.neutron [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Successfully created port: c09a7028-3a46-4c8a-8d63-5c4157aaaa9c {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 868.352871] env[63355]: INFO nova.scheduler.client.report [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Deleted allocations for instance dd80c3be-ec92-4a67-a373-3e2276752feb [ 868.520652] env[63355]: DEBUG nova.compute.manager [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 868.757182] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9370b75d-0bfa-4ca2-b696-b97137056c8d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.764554] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24e429c-2fe0-4dc2-8a12-62232e178f26 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.794272] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfba5423-cd69-4e22-9681-3ee2fd602fba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.801346] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47927a20-c1bf-448d-bbdd-8196e38c2778 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.819058] env[63355]: DEBUG nova.compute.provider_tree [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.863482] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9ac4bd18-39de-4360-80d7-f5afb21ce529 tempest-ServersAdminTestJSON-69647992 tempest-ServersAdminTestJSON-69647992-project-member] Lock "dd80c3be-ec92-4a67-a373-3e2276752feb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 188.805s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.322557] env[63355]: DEBUG nova.scheduler.client.report [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.366033] env[63355]: DEBUG nova.compute.manager [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 869.531758] env[63355]: DEBUG nova.compute.manager [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 869.555317] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.555513] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.555955] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.555955] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.556051] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.556421] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.556473] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.556626] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.556810] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.556975] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.557172] env[63355]: DEBUG nova.virt.hardware [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.558023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcb0a9e-ac8f-42cd-9a78-2625cc1576f9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.565960] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4f55c0-7fed-4f18-9178-d50f582d18b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.827562] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.828112] env[63355]: DEBUG nova.compute.manager [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 869.830692] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.997s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.891278] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.906798] env[63355]: DEBUG nova.compute.manager [req-c72cad04-e69e-4d6a-ad21-719f1f431154 req-d4a2d6c5-93ac-449b-99db-1ee443601526 service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Received event network-vif-plugged-c09a7028-3a46-4c8a-8d63-5c4157aaaa9c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.906895] env[63355]: DEBUG oslo_concurrency.lockutils [req-c72cad04-e69e-4d6a-ad21-719f1f431154 req-d4a2d6c5-93ac-449b-99db-1ee443601526 service nova] Acquiring lock "3d918cab-6ed6-4a37-a024-28e3db1b779c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.907129] env[63355]: DEBUG oslo_concurrency.lockutils [req-c72cad04-e69e-4d6a-ad21-719f1f431154 req-d4a2d6c5-93ac-449b-99db-1ee443601526 service nova] Lock "3d918cab-6ed6-4a37-a024-28e3db1b779c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.907365] env[63355]: DEBUG oslo_concurrency.lockutils [req-c72cad04-e69e-4d6a-ad21-719f1f431154 req-d4a2d6c5-93ac-449b-99db-1ee443601526 service nova] Lock "3d918cab-6ed6-4a37-a024-28e3db1b779c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.907486] env[63355]: DEBUG nova.compute.manager [req-c72cad04-e69e-4d6a-ad21-719f1f431154 req-d4a2d6c5-93ac-449b-99db-1ee443601526 service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] No waiting events found dispatching network-vif-plugged-c09a7028-3a46-4c8a-8d63-5c4157aaaa9c {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.907718] env[63355]: WARNING nova.compute.manager [req-c72cad04-e69e-4d6a-ad21-719f1f431154 req-d4a2d6c5-93ac-449b-99db-1ee443601526 service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Received unexpected event network-vif-plugged-c09a7028-3a46-4c8a-8d63-5c4157aaaa9c for instance with vm_state building and task_state spawning. [ 870.336119] env[63355]: DEBUG nova.compute.utils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 870.341672] env[63355]: DEBUG nova.compute.manager [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 870.342600] env[63355]: DEBUG nova.network.neutron [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 870.403888] env[63355]: DEBUG nova.network.neutron [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Successfully updated port: c09a7028-3a46-4c8a-8d63-5c4157aaaa9c {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.421854] env[63355]: DEBUG nova.policy [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3721c4a73df54ae9b44110cb9e8590a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc547065748241e8ac7b6c499ddaea66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.438120] env[63355]: DEBUG nova.compute.manager [req-48b3ca2b-72ed-4c9b-af77-85874e6ace48 req-9f0ef4f7-8c18-4c6c-8254-aab470b55ede service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Received event network-changed-c09a7028-3a46-4c8a-8d63-5c4157aaaa9c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.438325] env[63355]: DEBUG nova.compute.manager [req-48b3ca2b-72ed-4c9b-af77-85874e6ace48 req-9f0ef4f7-8c18-4c6c-8254-aab470b55ede service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Refreshing instance network info cache due to event network-changed-c09a7028-3a46-4c8a-8d63-5c4157aaaa9c. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.438536] env[63355]: DEBUG oslo_concurrency.lockutils [req-48b3ca2b-72ed-4c9b-af77-85874e6ace48 req-9f0ef4f7-8c18-4c6c-8254-aab470b55ede service nova] Acquiring lock "refresh_cache-3d918cab-6ed6-4a37-a024-28e3db1b779c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.438677] env[63355]: DEBUG oslo_concurrency.lockutils [req-48b3ca2b-72ed-4c9b-af77-85874e6ace48 req-9f0ef4f7-8c18-4c6c-8254-aab470b55ede service nova] Acquired lock "refresh_cache-3d918cab-6ed6-4a37-a024-28e3db1b779c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.438834] env[63355]: DEBUG nova.network.neutron [req-48b3ca2b-72ed-4c9b-af77-85874e6ace48 req-9f0ef4f7-8c18-4c6c-8254-aab470b55ede service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Refreshing network info cache for port c09a7028-3a46-4c8a-8d63-5c4157aaaa9c {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.619255] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2831b9-9502-448a-b260-055c4150b043 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.624645] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4106ac3-ecb4-4361-b3ec-6a1ba21f0d68 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.658665] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5bc371-9b5d-4fd4-b8d8-05fb0018d21d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.666044] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e144bf-64f6-479a-83d8-d2c28aaa0891 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.681245] env[63355]: DEBUG nova.compute.provider_tree [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.844631] env[63355]: DEBUG nova.compute.manager [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 870.906557] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquiring lock "refresh_cache-3d918cab-6ed6-4a37-a024-28e3db1b779c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.935777] env[63355]: DEBUG nova.network.neutron [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Successfully created port: 8fa3518a-34fb-4be4-992a-e08d9bfabe96 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 870.986528] env[63355]: DEBUG nova.network.neutron [req-48b3ca2b-72ed-4c9b-af77-85874e6ace48 req-9f0ef4f7-8c18-4c6c-8254-aab470b55ede service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.104736] env[63355]: DEBUG nova.network.neutron [req-48b3ca2b-72ed-4c9b-af77-85874e6ace48 req-9f0ef4f7-8c18-4c6c-8254-aab470b55ede service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.184516] env[63355]: DEBUG nova.scheduler.client.report [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.607517] env[63355]: DEBUG oslo_concurrency.lockutils [req-48b3ca2b-72ed-4c9b-af77-85874e6ace48 req-9f0ef4f7-8c18-4c6c-8254-aab470b55ede service nova] Releasing lock "refresh_cache-3d918cab-6ed6-4a37-a024-28e3db1b779c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.609561] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquired lock "refresh_cache-3d918cab-6ed6-4a37-a024-28e3db1b779c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.609561] env[63355]: DEBUG nova.network.neutron [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.690624] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.860s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.691244] env[63355]: ERROR nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e7c1c9cf-6026-408e-8353-37aeb78418d7, please check neutron logs for more information. [ 871.691244] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Traceback (most recent call last): [ 871.691244] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 871.691244] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] self.driver.spawn(context, instance, image_meta, [ 871.691244] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 871.691244] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 871.691244] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 871.691244] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] vm_ref = self.build_virtual_machine(instance, [ 871.691244] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 871.691244] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 871.691244] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] for vif in network_info: [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] return self._sync_wrapper(fn, *args, **kwargs) [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] self.wait() [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] self[:] = self._gt.wait() [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] return self._exit_event.wait() [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] current.throw(*self._exc) [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 871.691549] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] result = function(*args, **kwargs) [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] return func(*args, **kwargs) [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] raise e [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] nwinfo = self.network_api.allocate_for_instance( [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] created_port_ids = self._update_ports_for_instance( [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] with excutils.save_and_reraise_exception(): [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] self.force_reraise() [ 871.691976] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 871.692346] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] raise self.value [ 871.692346] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 871.692346] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] updated_port = self._update_port( [ 871.692346] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 871.692346] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] _ensure_no_port_binding_failure(port) [ 871.692346] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 871.692346] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] raise exception.PortBindingFailed(port_id=port['id']) [ 871.692346] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] nova.exception.PortBindingFailed: Binding failed for port e7c1c9cf-6026-408e-8353-37aeb78418d7, please check neutron logs for more information. [ 871.692346] env[63355]: ERROR nova.compute.manager [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] [ 871.692346] env[63355]: DEBUG nova.compute.utils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Binding failed for port e7c1c9cf-6026-408e-8353-37aeb78418d7, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 871.693261] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.978s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.695037] env[63355]: INFO nova.compute.claims [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.698026] env[63355]: DEBUG nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Build of instance 8a44bcfb-406c-45f6-b027-86d8980bc5d1 was re-scheduled: Binding failed for port e7c1c9cf-6026-408e-8353-37aeb78418d7, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 871.698401] env[63355]: DEBUG nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 871.698617] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "refresh_cache-8a44bcfb-406c-45f6-b027-86d8980bc5d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.698759] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired lock "refresh_cache-8a44bcfb-406c-45f6-b027-86d8980bc5d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.698914] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.858838] env[63355]: DEBUG nova.compute.manager [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 871.891090] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 871.891090] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 871.891280] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.891420] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 871.891572] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.891714] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 871.891915] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 871.892172] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 871.892360] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 871.893291] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 871.893291] env[63355]: DEBUG nova.virt.hardware [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 871.895415] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42556193-bc7a-4574-9776-abb8469d9be4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.901618] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d74967d5-7ae4-439c-ac5b-fc45022ac18d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.150902] env[63355]: DEBUG nova.network.neutron [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.230893] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.434035] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.461870] env[63355]: DEBUG nova.network.neutron [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Updating instance_info_cache with network_info: [{"id": "c09a7028-3a46-4c8a-8d63-5c4157aaaa9c", "address": "fa:16:3e:71:73:b8", "network": {"id": "fc709451-70bf-4a7e-84b3-71b7147e08db", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-764229386-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a26599cf04b24e17aac6f682b641c703", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc09a7028-3a", "ovs_interfaceid": "c09a7028-3a46-4c8a-8d63-5c4157aaaa9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.839356] env[63355]: DEBUG nova.compute.manager [req-cb4fad0c-8d73-48d8-93d1-3b5a20a037c9 req-bf864549-e89f-4bba-877e-7eb1c758a72d service nova] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Received event network-vif-plugged-8fa3518a-34fb-4be4-992a-e08d9bfabe96 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.839920] env[63355]: DEBUG oslo_concurrency.lockutils [req-cb4fad0c-8d73-48d8-93d1-3b5a20a037c9 req-bf864549-e89f-4bba-877e-7eb1c758a72d service nova] Acquiring lock "5591bb55-83d7-4301-a3f9-fde945632344-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.839920] env[63355]: DEBUG oslo_concurrency.lockutils [req-cb4fad0c-8d73-48d8-93d1-3b5a20a037c9 req-bf864549-e89f-4bba-877e-7eb1c758a72d service nova] Lock "5591bb55-83d7-4301-a3f9-fde945632344-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.840099] env[63355]: DEBUG oslo_concurrency.lockutils [req-cb4fad0c-8d73-48d8-93d1-3b5a20a037c9 req-bf864549-e89f-4bba-877e-7eb1c758a72d service nova] Lock "5591bb55-83d7-4301-a3f9-fde945632344-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.840197] env[63355]: DEBUG nova.compute.manager [req-cb4fad0c-8d73-48d8-93d1-3b5a20a037c9 req-bf864549-e89f-4bba-877e-7eb1c758a72d service nova] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] No waiting events found dispatching network-vif-plugged-8fa3518a-34fb-4be4-992a-e08d9bfabe96 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 872.840355] env[63355]: WARNING nova.compute.manager [req-cb4fad0c-8d73-48d8-93d1-3b5a20a037c9 req-bf864549-e89f-4bba-877e-7eb1c758a72d service nova] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Received unexpected event network-vif-plugged-8fa3518a-34fb-4be4-992a-e08d9bfabe96 for instance with vm_state building and task_state spawning. [ 872.937090] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Releasing lock "refresh_cache-8a44bcfb-406c-45f6-b027-86d8980bc5d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.937468] env[63355]: DEBUG nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 872.937656] env[63355]: DEBUG nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 872.937821] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 872.956726] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.959093] env[63355]: DEBUG nova.network.neutron [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Successfully updated port: 8fa3518a-34fb-4be4-992a-e08d9bfabe96 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 872.964521] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Releasing lock "refresh_cache-3d918cab-6ed6-4a37-a024-28e3db1b779c" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.964891] env[63355]: DEBUG nova.compute.manager [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Instance network_info: |[{"id": "c09a7028-3a46-4c8a-8d63-5c4157aaaa9c", "address": "fa:16:3e:71:73:b8", "network": {"id": "fc709451-70bf-4a7e-84b3-71b7147e08db", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-764229386-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a26599cf04b24e17aac6f682b641c703", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc09a7028-3a", "ovs_interfaceid": "c09a7028-3a46-4c8a-8d63-5c4157aaaa9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.965584] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:73:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c20f5114-0866-45b3-9a7c-62f113ff83fa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c09a7028-3a46-4c8a-8d63-5c4157aaaa9c', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.978073] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Creating folder: Project (a26599cf04b24e17aac6f682b641c703). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.979734] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-647e17fe-d3f7-4751-a308-77f2a14949d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.994024] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Created folder: Project (a26599cf04b24e17aac6f682b641c703) in parent group-v287607. [ 872.994279] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Creating folder: Instances. Parent ref: group-v287637. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.994494] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-30c8a8e8-7581-4ab0-a7d1-66644c8954f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.004921] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Created folder: Instances in parent group-v287637. [ 873.005191] env[63355]: DEBUG oslo.service.loopingcall [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.005659] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 873.005868] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8eadcd9c-ed97-4d8f-92e4-46e355ce0216 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.026311] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.026311] env[63355]: value = "task-1349594" [ 873.026311] env[63355]: _type = "Task" [ 873.026311] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.033807] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349594, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.071038] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df13969b-b101-4061-b825-35cf74af26ae {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.079198] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df96b6c1-9c88-42d8-a844-0b0cb8657501 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.112721] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e71471-2b5b-45fd-94c8-cf25d2e86672 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.120277] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4813681e-e2ef-4cff-8ada-01cafacf629c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.134614] env[63355]: DEBUG nova.compute.provider_tree [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.461867] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "refresh_cache-5591bb55-83d7-4301-a3f9-fde945632344" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.462471] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "refresh_cache-5591bb55-83d7-4301-a3f9-fde945632344" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.462471] env[63355]: DEBUG nova.network.neutron [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 873.463733] env[63355]: DEBUG nova.network.neutron [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.535837] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349594, 'name': CreateVM_Task, 'duration_secs': 0.296388} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.536012] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.542725] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.542898] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.543564] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.543564] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-199707be-6c30-4b2a-9794-bb7303f333af {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.548149] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for the task: (returnval){ [ 873.548149] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52063555-b880-32c3-fa7e-3b68084fe6cf" [ 873.548149] env[63355]: _type = "Task" [ 873.548149] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.557302] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52063555-b880-32c3-fa7e-3b68084fe6cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.637906] env[63355]: DEBUG nova.scheduler.client.report [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.969480] env[63355]: INFO nova.compute.manager [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 8a44bcfb-406c-45f6-b027-86d8980bc5d1] Took 1.03 seconds to deallocate network for instance. [ 874.026063] env[63355]: DEBUG nova.network.neutron [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.058724] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52063555-b880-32c3-fa7e-3b68084fe6cf, 'name': SearchDatastore_Task, 'duration_secs': 0.008711} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.059128] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.059430] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.059706] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.059911] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.060162] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.060456] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a28f313f-dd45-4cc2-aed6-a3d32eb93c39 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.067970] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.068200] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.068934] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-797123eb-db87-48e0-a118-e3c73048cd7c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.074269] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for the task: (returnval){ [ 874.074269] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b1f03e-38a1-c6fd-8391-b877aeb890d9" [ 874.074269] env[63355]: _type = "Task" [ 874.074269] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.082051] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b1f03e-38a1-c6fd-8391-b877aeb890d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.142648] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.143255] env[63355]: DEBUG nova.compute.manager [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.146022] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.824s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.146217] env[63355]: DEBUG nova.objects.instance [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63355) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 874.235583] env[63355]: DEBUG nova.network.neutron [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Updating instance_info_cache with network_info: [{"id": "8fa3518a-34fb-4be4-992a-e08d9bfabe96", "address": "fa:16:3e:9a:cc:79", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fa3518a-34", "ovs_interfaceid": "8fa3518a-34fb-4be4-992a-e08d9bfabe96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.587645] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b1f03e-38a1-c6fd-8391-b877aeb890d9, 'name': SearchDatastore_Task, 'duration_secs': 0.00792} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.588744] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f422d04-2bb7-40cc-899e-e2699e636f1d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.595188] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for the task: (returnval){ [ 874.595188] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f6aa98-9b91-3bc1-62d4-fdcf8c3d22ab" [ 874.595188] env[63355]: _type = "Task" [ 874.595188] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.602932] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f6aa98-9b91-3bc1-62d4-fdcf8c3d22ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.652400] env[63355]: DEBUG nova.compute.utils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.657421] env[63355]: DEBUG nova.compute.manager [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.657691] env[63355]: DEBUG nova.network.neutron [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 874.695252] env[63355]: DEBUG nova.policy [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77243643ea724b72858a8682a2a054a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d47b671ea9c429391cbdae7e24adadf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.739197] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "refresh_cache-5591bb55-83d7-4301-a3f9-fde945632344" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.739361] env[63355]: DEBUG nova.compute.manager [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Instance network_info: |[{"id": "8fa3518a-34fb-4be4-992a-e08d9bfabe96", "address": "fa:16:3e:9a:cc:79", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fa3518a-34", "ovs_interfaceid": "8fa3518a-34fb-4be4-992a-e08d9bfabe96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 874.740261] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:cc:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8fa3518a-34fb-4be4-992a-e08d9bfabe96', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 874.747769] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Creating folder: Project (bc547065748241e8ac7b6c499ddaea66). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 874.750674] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7cd4c53d-ae75-4f87-b287-00704f0c7c54 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.761189] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Created folder: Project (bc547065748241e8ac7b6c499ddaea66) in parent group-v287607. [ 874.761424] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Creating folder: Instances. Parent ref: group-v287640. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 874.761673] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ef688c1-c5b3-4b11-8421-dc02fd968786 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.770235] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Created folder: Instances in parent group-v287640. [ 874.770466] env[63355]: DEBUG oslo.service.loopingcall [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.770650] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 874.770841] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae7fd88e-0fde-4fe6-8a28-36e140aeaca6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.788900] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 874.788900] env[63355]: value = "task-1349597" [ 874.788900] env[63355]: _type = "Task" [ 874.788900] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.796319] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349597, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.869558] env[63355]: DEBUG nova.compute.manager [req-8c72efa4-3796-47a0-8b3b-cfde0af99e76 req-28f354c0-a70d-44f9-b207-fde48cecd570 service nova] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Received event network-changed-8fa3518a-34fb-4be4-992a-e08d9bfabe96 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.869776] env[63355]: DEBUG nova.compute.manager [req-8c72efa4-3796-47a0-8b3b-cfde0af99e76 req-28f354c0-a70d-44f9-b207-fde48cecd570 service nova] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Refreshing instance network info cache due to event network-changed-8fa3518a-34fb-4be4-992a-e08d9bfabe96. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 874.869985] env[63355]: DEBUG oslo_concurrency.lockutils [req-8c72efa4-3796-47a0-8b3b-cfde0af99e76 req-28f354c0-a70d-44f9-b207-fde48cecd570 service nova] Acquiring lock "refresh_cache-5591bb55-83d7-4301-a3f9-fde945632344" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.870459] env[63355]: DEBUG oslo_concurrency.lockutils [req-8c72efa4-3796-47a0-8b3b-cfde0af99e76 req-28f354c0-a70d-44f9-b207-fde48cecd570 service nova] Acquired lock "refresh_cache-5591bb55-83d7-4301-a3f9-fde945632344" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.870704] env[63355]: DEBUG nova.network.neutron [req-8c72efa4-3796-47a0-8b3b-cfde0af99e76 req-28f354c0-a70d-44f9-b207-fde48cecd570 service nova] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Refreshing network info cache for port 8fa3518a-34fb-4be4-992a-e08d9bfabe96 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.020909] env[63355]: INFO nova.scheduler.client.report [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Deleted allocations for instance 8a44bcfb-406c-45f6-b027-86d8980bc5d1 [ 875.084356] env[63355]: DEBUG nova.network.neutron [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Successfully created port: 2348dcf0-b9a5-491f-87cb-a14e9e002368 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.108066] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f6aa98-9b91-3bc1-62d4-fdcf8c3d22ab, 'name': SearchDatastore_Task, 'duration_secs': 0.014349} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.108235] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.108523] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 3d918cab-6ed6-4a37-a024-28e3db1b779c/3d918cab-6ed6-4a37-a024-28e3db1b779c.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 875.109400] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1bc7fc86-4289-4cdb-97f5-9110cae821bc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.118485] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for the task: (returnval){ [ 875.118485] env[63355]: value = "task-1349598" [ 875.118485] env[63355]: _type = "Task" [ 875.118485] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.127502] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349598, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.158380] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f6f3614c-0da8-4539-825e-ef886aa4b44b tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.160466] env[63355]: DEBUG nova.compute.manager [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.163896] env[63355]: DEBUG oslo_concurrency.lockutils [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.609s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.164066] env[63355]: DEBUG nova.objects.instance [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63355) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 875.308542] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349597, 'name': CreateVM_Task, 'duration_secs': 0.273436} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.308916] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.310039] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.310039] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.310305] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.310622] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-029a5c86-3fd2-4cf6-a35f-d79d8ee11239 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.316728] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 875.316728] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ade0ed-36e6-18a3-d3ae-c61927d9be48" [ 875.316728] env[63355]: _type = "Task" [ 875.316728] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.327969] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ade0ed-36e6-18a3-d3ae-c61927d9be48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.531013] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97d4ff09-32de-4bfd-a7bc-a1b0300bbcec tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "8a44bcfb-406c-45f6-b027-86d8980bc5d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 183.956s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.628207] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349598, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493836} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.628490] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 3d918cab-6ed6-4a37-a024-28e3db1b779c/3d918cab-6ed6-4a37-a024-28e3db1b779c.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.628715] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.628963] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4bf6a381-6789-493c-b327-98f7a46f8d7d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.635847] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for the task: (returnval){ [ 875.635847] env[63355]: value = "task-1349599" [ 875.635847] env[63355]: _type = "Task" [ 875.635847] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.643524] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349599, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.749098] env[63355]: DEBUG nova.network.neutron [req-8c72efa4-3796-47a0-8b3b-cfde0af99e76 req-28f354c0-a70d-44f9-b207-fde48cecd570 service nova] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Updated VIF entry in instance network info cache for port 8fa3518a-34fb-4be4-992a-e08d9bfabe96. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 875.749523] env[63355]: DEBUG nova.network.neutron [req-8c72efa4-3796-47a0-8b3b-cfde0af99e76 req-28f354c0-a70d-44f9-b207-fde48cecd570 service nova] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Updating instance_info_cache with network_info: [{"id": "8fa3518a-34fb-4be4-992a-e08d9bfabe96", "address": "fa:16:3e:9a:cc:79", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fa3518a-34", "ovs_interfaceid": "8fa3518a-34fb-4be4-992a-e08d9bfabe96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.830993] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ade0ed-36e6-18a3-d3ae-c61927d9be48, 'name': SearchDatastore_Task, 'duration_secs': 0.054672} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.831373] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.831760] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 875.832264] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.832264] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.832457] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 875.832756] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1faaa4b3-0e73-4674-b15a-00bdcf3aef00 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.840885] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 875.841126] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 875.841856] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b98f4fa7-3306-4c80-addf-1d5df3601d18 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.848684] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 875.848684] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52848da1-bc28-fb35-7f54-5c8d86d5feaa" [ 875.848684] env[63355]: _type = "Task" [ 875.848684] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.856467] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52848da1-bc28-fb35-7f54-5c8d86d5feaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.034905] env[63355]: DEBUG nova.compute.manager [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 876.147975] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349599, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066331} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.148501] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.149696] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e2bad6-28ec-4298-9ff7-12c95335fc08 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.177184] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 3d918cab-6ed6-4a37-a024-28e3db1b779c/3d918cab-6ed6-4a37-a024-28e3db1b779c.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.178640] env[63355]: DEBUG nova.compute.manager [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.181013] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-570247a7-9add-4fa7-a070-749a62735125 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.202699] env[63355]: DEBUG oslo_concurrency.lockutils [None req-21ff5d7f-4ea0-4bff-ae59-2fed17233c49 tempest-ServersAdmin275Test-2119330454 tempest-ServersAdmin275Test-2119330454-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.038s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.204114] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.722s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.208284] env[63355]: INFO nova.compute.claims [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.223860] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for the task: (returnval){ [ 876.223860] env[63355]: value = "task-1349600" [ 876.223860] env[63355]: _type = "Task" [ 876.223860] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.234588] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349600, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.252303] env[63355]: DEBUG oslo_concurrency.lockutils [req-8c72efa4-3796-47a0-8b3b-cfde0af99e76 req-28f354c0-a70d-44f9-b207-fde48cecd570 service nova] Releasing lock "refresh_cache-5591bb55-83d7-4301-a3f9-fde945632344" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.264657] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.265230] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.265461] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.265668] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.265841] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.265974] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.266200] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.266354] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.266529] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.267211] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.267445] env[63355]: DEBUG nova.virt.hardware [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.268327] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6180d73-fc04-4a77-9429-fa267cb5587e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.277751] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d49a64f-bde9-4534-a1df-c6364333dd48 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.360961] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52848da1-bc28-fb35-7f54-5c8d86d5feaa, 'name': SearchDatastore_Task, 'duration_secs': 0.007864} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.361979] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f71bc2d-bbdf-42fa-af5d-c69a90ec5368 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.367544] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 876.367544] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f6c0f0-1d6a-886e-f45e-0f61ecda6ef7" [ 876.367544] env[63355]: _type = "Task" [ 876.367544] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.375106] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f6c0f0-1d6a-886e-f45e-0f61ecda6ef7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.569246] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.735105] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349600, 'name': ReconfigVM_Task, 'duration_secs': 0.313071} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.735105] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 3d918cab-6ed6-4a37-a024-28e3db1b779c/3d918cab-6ed6-4a37-a024-28e3db1b779c.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.735900] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d738827-13ca-44ab-91cf-323e0c464ccf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.746019] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for the task: (returnval){ [ 876.746019] env[63355]: value = "task-1349601" [ 876.746019] env[63355]: _type = "Task" [ 876.746019] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.753698] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349601, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.878164] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f6c0f0-1d6a-886e-f45e-0f61ecda6ef7, 'name': SearchDatastore_Task, 'duration_secs': 0.01019} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.878735] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.878735] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 5591bb55-83d7-4301-a3f9-fde945632344/5591bb55-83d7-4301-a3f9-fde945632344.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 876.878954] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-753b2910-db0f-4097-ac70-60058d9531e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.890190] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 876.890190] env[63355]: value = "task-1349602" [ 876.890190] env[63355]: _type = "Task" [ 876.890190] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.899825] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349602, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.128032] env[63355]: DEBUG nova.network.neutron [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Successfully updated port: 2348dcf0-b9a5-491f-87cb-a14e9e002368 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.263322] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349601, 'name': Rename_Task, 'duration_secs': 0.142831} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.266583] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 877.268177] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f90f8cd-8258-433b-ab7e-feda800cb8b3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.275236] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for the task: (returnval){ [ 877.275236] env[63355]: value = "task-1349603" [ 877.275236] env[63355]: _type = "Task" [ 877.275236] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.283761] env[63355]: DEBUG nova.compute.manager [req-885d64d7-a25d-4684-946b-08bde160e4e2 req-1e96aa72-3b45-4369-a65e-e73f0457d476 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received event network-vif-plugged-2348dcf0-b9a5-491f-87cb-a14e9e002368 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.284080] env[63355]: DEBUG oslo_concurrency.lockutils [req-885d64d7-a25d-4684-946b-08bde160e4e2 req-1e96aa72-3b45-4369-a65e-e73f0457d476 service nova] Acquiring lock "386e847e-967b-4247-9730-cdc5ac251474-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.284170] env[63355]: DEBUG oslo_concurrency.lockutils [req-885d64d7-a25d-4684-946b-08bde160e4e2 req-1e96aa72-3b45-4369-a65e-e73f0457d476 service nova] Lock "386e847e-967b-4247-9730-cdc5ac251474-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.284342] env[63355]: DEBUG oslo_concurrency.lockutils [req-885d64d7-a25d-4684-946b-08bde160e4e2 req-1e96aa72-3b45-4369-a65e-e73f0457d476 service nova] Lock "386e847e-967b-4247-9730-cdc5ac251474-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.284504] env[63355]: DEBUG nova.compute.manager [req-885d64d7-a25d-4684-946b-08bde160e4e2 req-1e96aa72-3b45-4369-a65e-e73f0457d476 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] No waiting events found dispatching network-vif-plugged-2348dcf0-b9a5-491f-87cb-a14e9e002368 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.284669] env[63355]: WARNING nova.compute.manager [req-885d64d7-a25d-4684-946b-08bde160e4e2 req-1e96aa72-3b45-4369-a65e-e73f0457d476 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received unexpected event network-vif-plugged-2348dcf0-b9a5-491f-87cb-a14e9e002368 for instance with vm_state building and task_state spawning. [ 877.292510] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349603, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.400426] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349602, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50083} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.402891] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 5591bb55-83d7-4301-a3f9-fde945632344/5591bb55-83d7-4301-a3f9-fde945632344.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.403138] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.403651] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-994d4234-a5d1-4212-b583-21ea0fad41e3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.409484] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 877.409484] env[63355]: value = "task-1349604" [ 877.409484] env[63355]: _type = "Task" [ 877.409484] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.419235] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349604, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.556551] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ea47a4-11f7-4d19-9ff6-10ec5eb419c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.564635] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981ba29e-f2cb-42a5-9498-384e33bc24e8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.605595] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f08d5a8-0d74-4f6c-b8b1-600b9a8c6fc1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.611882] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquiring lock "a408e19a-0960-430d-8550-0a304c63da61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.612140] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Lock "a408e19a-0960-430d-8550-0a304c63da61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.618228] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed18414-365f-4fe8-a374-ed99892ac867 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.634347] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.634492] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.634641] env[63355]: DEBUG nova.network.neutron [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.639544] env[63355]: DEBUG nova.compute.provider_tree [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.786287] env[63355]: DEBUG oslo_vmware.api [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349603, 'name': PowerOnVM_Task, 'duration_secs': 0.502485} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.786623] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.786828] env[63355]: INFO nova.compute.manager [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Took 8.25 seconds to spawn the instance on the hypervisor. [ 877.787010] env[63355]: DEBUG nova.compute.manager [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.787799] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18039c6f-96b0-4d46-8d2e-6e9d828770fc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.920080] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349604, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069664} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.920401] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.921216] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496ca0bd-3583-4f7f-9dae-a0158ec9d18c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.943609] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 5591bb55-83d7-4301-a3f9-fde945632344/5591bb55-83d7-4301-a3f9-fde945632344.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.943907] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a7b2b06-2170-4213-b054-f287374ff44e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.965008] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 877.965008] env[63355]: value = "task-1349605" [ 877.965008] env[63355]: _type = "Task" [ 877.965008] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.973904] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349605, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.142175] env[63355]: DEBUG nova.scheduler.client.report [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.201954] env[63355]: DEBUG nova.network.neutron [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.310807] env[63355]: INFO nova.compute.manager [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Took 39.17 seconds to build instance. [ 878.424397] env[63355]: DEBUG nova.network.neutron [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [{"id": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "address": "fa:16:3e:a9:13:18", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348dcf0-b9", "ovs_interfaceid": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.438500] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.438708] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.480362] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349605, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.651019] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.651019] env[63355]: DEBUG nova.compute.manager [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 878.652048] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.897s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.652828] env[63355]: DEBUG nova.objects.instance [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lazy-loading 'resources' on Instance uuid 9989c129-07d3-4af9-9a86-1e2746d6ac70 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.812040] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8de97b51-3073-42ed-a41d-eb295d47316c tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lock "3d918cab-6ed6-4a37-a024-28e3db1b779c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 172.998s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.927451] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.927858] env[63355]: DEBUG nova.compute.manager [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Instance network_info: |[{"id": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "address": "fa:16:3e:a9:13:18", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348dcf0-b9", "ovs_interfaceid": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.928320] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:13:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e41070eb-3ac1-4ca9-a3d0-fd65893a97de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2348dcf0-b9a5-491f-87cb-a14e9e002368', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.935990] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Creating folder: Project (4d47b671ea9c429391cbdae7e24adadf). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.936325] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d01c3a4c-4123-4f4e-86cd-ed39d3b581c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.946151] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Created folder: Project (4d47b671ea9c429391cbdae7e24adadf) in parent group-v287607. [ 878.946362] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Creating folder: Instances. Parent ref: group-v287643. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.946828] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-82c057dd-de38-4de2-a943-d2f15be36a1b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.956469] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Created folder: Instances in parent group-v287643. [ 878.956755] env[63355]: DEBUG oslo.service.loopingcall [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.957587] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 878.957821] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d99a11d0-b2cd-4cac-ba2a-a6c69e370d45 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.982023] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349605, 'name': ReconfigVM_Task, 'duration_secs': 0.956272} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.983223] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 5591bb55-83d7-4301-a3f9-fde945632344/5591bb55-83d7-4301-a3f9-fde945632344.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.983839] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.983839] env[63355]: value = "task-1349608" [ 878.983839] env[63355]: _type = "Task" [ 878.983839] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.984045] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ae8209d-f5d1-4615-b82d-ecb28328bd14 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.994468] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349608, 'name': CreateVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.994656] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 878.994656] env[63355]: value = "task-1349609" [ 878.994656] env[63355]: _type = "Task" [ 878.994656] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.001862] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349609, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.156450] env[63355]: DEBUG nova.compute.utils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.162357] env[63355]: DEBUG nova.compute.manager [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.164085] env[63355]: DEBUG nova.network.neutron [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 879.249164] env[63355]: DEBUG nova.policy [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4ec26bc0e3049b39a8e2c901e7b4d81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1cd345e63584e108f502bb46ec3f02a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 879.314775] env[63355]: DEBUG nova.compute.manager [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.367887] env[63355]: DEBUG nova.compute.manager [req-ac489b1c-4114-40a0-a03c-8389b945de6d req-5fafbe4d-b640-4a10-b2b1-efd53db08e09 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received event network-changed-2348dcf0-b9a5-491f-87cb-a14e9e002368 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.368167] env[63355]: DEBUG nova.compute.manager [req-ac489b1c-4114-40a0-a03c-8389b945de6d req-5fafbe4d-b640-4a10-b2b1-efd53db08e09 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Refreshing instance network info cache due to event network-changed-2348dcf0-b9a5-491f-87cb-a14e9e002368. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 879.368682] env[63355]: DEBUG oslo_concurrency.lockutils [req-ac489b1c-4114-40a0-a03c-8389b945de6d req-5fafbe4d-b640-4a10-b2b1-efd53db08e09 service nova] Acquiring lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.368840] env[63355]: DEBUG oslo_concurrency.lockutils [req-ac489b1c-4114-40a0-a03c-8389b945de6d req-5fafbe4d-b640-4a10-b2b1-efd53db08e09 service nova] Acquired lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.369014] env[63355]: DEBUG nova.network.neutron [req-ac489b1c-4114-40a0-a03c-8389b945de6d req-5fafbe4d-b640-4a10-b2b1-efd53db08e09 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Refreshing network info cache for port 2348dcf0-b9a5-491f-87cb-a14e9e002368 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.460914] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ea8d73-861e-43b1-adf1-142d7c394a88 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.469733] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8a7311-fa64-4591-a58a-648c5a842632 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.506709] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfab75b-a243-4ca3-a64e-9272760055ca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.514774] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349608, 'name': CreateVM_Task, 'duration_secs': 0.304754} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.519435] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 879.519721] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349609, 'name': Rename_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.520711] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.520870] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.521189] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.521435] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d94e6f99-b4f7-4ee7-ad76-bfb368e77ce4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.525895] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3350349-a937-489c-af15-7388c85c8333 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.530460] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 879.530460] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52eed52d-db73-3b82-6bc9-5c5b72246309" [ 879.530460] env[63355]: _type = "Task" [ 879.530460] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.541608] env[63355]: DEBUG nova.compute.provider_tree [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.548159] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52eed52d-db73-3b82-6bc9-5c5b72246309, 'name': SearchDatastore_Task, 'duration_secs': 0.009061} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.548673] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.548673] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.548876] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.549063] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.549215] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.549460] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f04ad42c-0bef-4b0a-a918-98e16762fcfe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.557207] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.557386] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 879.558302] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee364d25-c3e0-4600-85b5-5b8a8e717d75 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.564208] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 879.564208] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c38dd1-d0ac-5cef-db4c-1abfc286f8df" [ 879.564208] env[63355]: _type = "Task" [ 879.564208] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.571962] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c38dd1-d0ac-5cef-db4c-1abfc286f8df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.663085] env[63355]: DEBUG nova.compute.manager [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 879.731787] env[63355]: DEBUG nova.network.neutron [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Successfully created port: 86bfab04-fd0a-4339-922e-241bb081e78e {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 879.836558] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.010432] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349609, 'name': Rename_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.049161] env[63355]: DEBUG nova.scheduler.client.report [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.075715] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c38dd1-d0ac-5cef-db4c-1abfc286f8df, 'name': SearchDatastore_Task, 'duration_secs': 0.019569} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.077143] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b05f4d8-a074-4f68-87bf-64efed8d01a6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.083404] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 880.083404] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529a6e38-9369-079e-1964-fe4082646901" [ 880.083404] env[63355]: _type = "Task" [ 880.083404] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.094034] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529a6e38-9369-079e-1964-fe4082646901, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.106155] env[63355]: DEBUG nova.network.neutron [req-ac489b1c-4114-40a0-a03c-8389b945de6d req-5fafbe4d-b640-4a10-b2b1-efd53db08e09 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updated VIF entry in instance network info cache for port 2348dcf0-b9a5-491f-87cb-a14e9e002368. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.106618] env[63355]: DEBUG nova.network.neutron [req-ac489b1c-4114-40a0-a03c-8389b945de6d req-5fafbe4d-b640-4a10-b2b1-efd53db08e09 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [{"id": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "address": "fa:16:3e:a9:13:18", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348dcf0-b9", "ovs_interfaceid": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.300690] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquiring lock "3d918cab-6ed6-4a37-a024-28e3db1b779c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.300986] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lock "3d918cab-6ed6-4a37-a024-28e3db1b779c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.301186] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquiring lock "3d918cab-6ed6-4a37-a024-28e3db1b779c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.301361] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lock "3d918cab-6ed6-4a37-a024-28e3db1b779c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.301527] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lock "3d918cab-6ed6-4a37-a024-28e3db1b779c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.303902] env[63355]: INFO nova.compute.manager [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Terminating instance [ 880.305759] env[63355]: DEBUG nova.compute.manager [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 880.305932] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.306791] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe072ec-bdd6-492d-82f2-cdee33fea913 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.314448] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.315038] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-024a151e-d762-4b75-941f-eef7a1aab091 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.321133] env[63355]: DEBUG oslo_vmware.api [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for the task: (returnval){ [ 880.321133] env[63355]: value = "task-1349610" [ 880.321133] env[63355]: _type = "Task" [ 880.321133] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.328916] env[63355]: DEBUG oslo_vmware.api [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349610, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.510715] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349609, 'name': Rename_Task, 'duration_secs': 1.134932} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.510971] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.511299] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3c09459-800a-4c06-a25f-6cc4007b67b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.517425] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 880.517425] env[63355]: value = "task-1349611" [ 880.517425] env[63355]: _type = "Task" [ 880.517425] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.526466] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349611, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.551476] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.899s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.553785] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.427s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.555263] env[63355]: INFO nova.compute.claims [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.583695] env[63355]: INFO nova.scheduler.client.report [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Deleted allocations for instance 9989c129-07d3-4af9-9a86-1e2746d6ac70 [ 880.597029] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529a6e38-9369-079e-1964-fe4082646901, 'name': SearchDatastore_Task, 'duration_secs': 0.013841} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.597510] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.597768] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 386e847e-967b-4247-9730-cdc5ac251474/386e847e-967b-4247-9730-cdc5ac251474.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.598051] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fd5ca2b-8ef9-43cb-a0a4-7054b3ccb200 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.604559] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 880.604559] env[63355]: value = "task-1349612" [ 880.604559] env[63355]: _type = "Task" [ 880.604559] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.608529] env[63355]: DEBUG oslo_concurrency.lockutils [req-ac489b1c-4114-40a0-a03c-8389b945de6d req-5fafbe4d-b640-4a10-b2b1-efd53db08e09 service nova] Releasing lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.613355] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349612, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.672939] env[63355]: DEBUG nova.compute.manager [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 880.709318] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.709687] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.709950] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.710318] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.710903] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.710903] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.711171] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.711558] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.711718] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.711996] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.712306] env[63355]: DEBUG nova.virt.hardware [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.713674] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604b23db-6983-47d4-93b8-46bfdf25a031 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.725091] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e95c4aa-ab7c-411d-9e21-158e8075300f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.834283] env[63355]: DEBUG oslo_vmware.api [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349610, 'name': PowerOffVM_Task, 'duration_secs': 0.212818} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.834622] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 880.834794] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 880.835104] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5108c2b-1034-4c76-8592-bae4373b1145 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.927771] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 880.928159] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 880.928381] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Deleting the datastore file [datastore2] 3d918cab-6ed6-4a37-a024-28e3db1b779c {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.928640] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d0ec7892-c94b-4080-9ca2-d37e0459051f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.937779] env[63355]: DEBUG oslo_vmware.api [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for the task: (returnval){ [ 880.937779] env[63355]: value = "task-1349614" [ 880.937779] env[63355]: _type = "Task" [ 880.937779] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.950057] env[63355]: DEBUG oslo_vmware.api [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349614, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.030556] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349611, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.096033] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c294ea6c-4849-47ee-b091-ece1cebb56a4 tempest-ServersAdmin275Test-1572014996 tempest-ServersAdmin275Test-1572014996-project-member] Lock "9989c129-07d3-4af9-9a86-1e2746d6ac70" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.053s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.118498] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349612, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.454939] env[63355]: DEBUG oslo_vmware.api [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Task: {'id': task-1349614, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.396916} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.455471] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 881.455799] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 881.456135] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 881.456446] env[63355]: INFO nova.compute.manager [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 881.456836] env[63355]: DEBUG oslo.service.loopingcall [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.457180] env[63355]: DEBUG nova.compute.manager [-] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 881.457344] env[63355]: DEBUG nova.network.neutron [-] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 881.528744] env[63355]: DEBUG oslo_vmware.api [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349611, 'name': PowerOnVM_Task, 'duration_secs': 0.808427} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.529027] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 881.529271] env[63355]: INFO nova.compute.manager [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Took 9.67 seconds to spawn the instance on the hypervisor. [ 881.529572] env[63355]: DEBUG nova.compute.manager [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.530440] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef85bc9c-201e-4618-89d1-8d1717e998a6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.569041] env[63355]: DEBUG nova.compute.manager [req-cd21a32a-3a76-4e03-bf0c-8a199da7c21a req-70c949b6-f718-4553-baec-1382f7b79b09 service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Received event network-vif-plugged-86bfab04-fd0a-4339-922e-241bb081e78e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.569271] env[63355]: DEBUG oslo_concurrency.lockutils [req-cd21a32a-3a76-4e03-bf0c-8a199da7c21a req-70c949b6-f718-4553-baec-1382f7b79b09 service nova] Acquiring lock "0b660647-697a-445b-bca4-82ef6975ea75-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.569470] env[63355]: DEBUG oslo_concurrency.lockutils [req-cd21a32a-3a76-4e03-bf0c-8a199da7c21a req-70c949b6-f718-4553-baec-1382f7b79b09 service nova] Lock "0b660647-697a-445b-bca4-82ef6975ea75-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.569636] env[63355]: DEBUG oslo_concurrency.lockutils [req-cd21a32a-3a76-4e03-bf0c-8a199da7c21a req-70c949b6-f718-4553-baec-1382f7b79b09 service nova] Lock "0b660647-697a-445b-bca4-82ef6975ea75-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.569803] env[63355]: DEBUG nova.compute.manager [req-cd21a32a-3a76-4e03-bf0c-8a199da7c21a req-70c949b6-f718-4553-baec-1382f7b79b09 service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] No waiting events found dispatching network-vif-plugged-86bfab04-fd0a-4339-922e-241bb081e78e {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.569961] env[63355]: WARNING nova.compute.manager [req-cd21a32a-3a76-4e03-bf0c-8a199da7c21a req-70c949b6-f718-4553-baec-1382f7b79b09 service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Received unexpected event network-vif-plugged-86bfab04-fd0a-4339-922e-241bb081e78e for instance with vm_state building and task_state spawning. [ 881.615821] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349612, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547697} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.618611] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 386e847e-967b-4247-9730-cdc5ac251474/386e847e-967b-4247-9730-cdc5ac251474.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.618891] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.620194] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc3d076f-03cd-40d0-9d1b-e04981372a47 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.634599] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 881.634599] env[63355]: value = "task-1349615" [ 881.634599] env[63355]: _type = "Task" [ 881.634599] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.645494] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349615, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.676088] env[63355]: DEBUG nova.network.neutron [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Successfully updated port: 86bfab04-fd0a-4339-922e-241bb081e78e {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.851042] env[63355]: DEBUG nova.compute.manager [req-3ad8253a-dfc4-4c2d-8042-5f12e52e9891 req-e38b9ad2-fb64-4fa7-83f9-37f65e9b67ff service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Received event network-vif-deleted-c09a7028-3a46-4c8a-8d63-5c4157aaaa9c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.851162] env[63355]: INFO nova.compute.manager [req-3ad8253a-dfc4-4c2d-8042-5f12e52e9891 req-e38b9ad2-fb64-4fa7-83f9-37f65e9b67ff service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Neutron deleted interface c09a7028-3a46-4c8a-8d63-5c4157aaaa9c; detaching it from the instance and deleting it from the info cache [ 881.851339] env[63355]: DEBUG nova.network.neutron [req-3ad8253a-dfc4-4c2d-8042-5f12e52e9891 req-e38b9ad2-fb64-4fa7-83f9-37f65e9b67ff service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.938021] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e77cbb-d65b-49c6-a46c-30754dc47e31 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.946217] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c36300-04ed-45e9-a54b-0f2d43abdc27 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.983938] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e4d524-8999-4277-9cd9-d0907bc7a8f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.993387] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af66b1cc-0212-4398-9539-d93a57b28d1e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.008353] env[63355]: DEBUG nova.compute.provider_tree [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.047853] env[63355]: INFO nova.compute.manager [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Took 40.60 seconds to build instance. [ 882.144282] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349615, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.366305} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.145092] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.145496] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dddbfd47-7084-4634-9fce-fb5bbe5edb12 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.167060] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 386e847e-967b-4247-9730-cdc5ac251474/386e847e-967b-4247-9730-cdc5ac251474.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.167341] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e141667-dc34-45e7-8e89-6f85cba9e930 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.186306] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "refresh_cache-0b660647-697a-445b-bca4-82ef6975ea75" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.186457] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquired lock "refresh_cache-0b660647-697a-445b-bca4-82ef6975ea75" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.186727] env[63355]: DEBUG nova.network.neutron [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.187874] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 882.187874] env[63355]: value = "task-1349616" [ 882.187874] env[63355]: _type = "Task" [ 882.187874] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.197086] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349616, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.328115] env[63355]: DEBUG nova.network.neutron [-] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.357299] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-193e60b9-28b6-4a6b-9206-809a2a6e9f84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.367145] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e5aaa6-0eb5-447a-b2a3-9dd95b29e4f3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.392308] env[63355]: DEBUG nova.compute.manager [req-3ad8253a-dfc4-4c2d-8042-5f12e52e9891 req-e38b9ad2-fb64-4fa7-83f9-37f65e9b67ff service nova] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Detach interface failed, port_id=c09a7028-3a46-4c8a-8d63-5c4157aaaa9c, reason: Instance 3d918cab-6ed6-4a37-a024-28e3db1b779c could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 882.511420] env[63355]: DEBUG nova.scheduler.client.report [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.551030] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f838392a-1b94-406b-8bd2-b1061ed22cd7 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "5591bb55-83d7-4301-a3f9-fde945632344" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 174.954s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.701784] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349616, 'name': ReconfigVM_Task, 'duration_secs': 0.338722} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.701784] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 386e847e-967b-4247-9730-cdc5ac251474/386e847e-967b-4247-9730-cdc5ac251474.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.702444] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1dfe0341-0a02-4c48-905c-a315c45dc15a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.709423] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 882.709423] env[63355]: value = "task-1349617" [ 882.709423] env[63355]: _type = "Task" [ 882.709423] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.717133] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349617, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.730346] env[63355]: DEBUG nova.network.neutron [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.830277] env[63355]: INFO nova.compute.manager [-] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Took 1.37 seconds to deallocate network for instance. [ 882.874210] env[63355]: DEBUG nova.network.neutron [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Updating instance_info_cache with network_info: [{"id": "86bfab04-fd0a-4339-922e-241bb081e78e", "address": "fa:16:3e:96:31:48", "network": {"id": "5930e4a2-0fe8-4b14-ad75-fca389ab1061", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.36", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfa2720fe39d48deb0cad174a7e5f3ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f39e3b37-7906-4bbc-820e-ceac74e4d827", "external-id": "nsx-vlan-transportzone-328", "segmentation_id": 328, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86bfab04-fd", "ovs_interfaceid": "86bfab04-fd0a-4339-922e-241bb081e78e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.016976] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.017518] env[63355]: DEBUG nova.compute.manager [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.020246] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.539s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.052708] env[63355]: DEBUG nova.compute.manager [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.219607] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349617, 'name': Rename_Task, 'duration_secs': 0.174357} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.219915] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.220190] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8fdc0bb-6628-401a-a595-6f780f4b37b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.227049] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 883.227049] env[63355]: value = "task-1349618" [ 883.227049] env[63355]: _type = "Task" [ 883.227049] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.234584] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349618, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.338570] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.376881] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Releasing lock "refresh_cache-0b660647-697a-445b-bca4-82ef6975ea75" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.377250] env[63355]: DEBUG nova.compute.manager [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Instance network_info: |[{"id": "86bfab04-fd0a-4339-922e-241bb081e78e", "address": "fa:16:3e:96:31:48", "network": {"id": "5930e4a2-0fe8-4b14-ad75-fca389ab1061", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.36", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfa2720fe39d48deb0cad174a7e5f3ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f39e3b37-7906-4bbc-820e-ceac74e4d827", "external-id": "nsx-vlan-transportzone-328", "segmentation_id": 328, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86bfab04-fd", "ovs_interfaceid": "86bfab04-fd0a-4339-922e-241bb081e78e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.377680] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:31:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f39e3b37-7906-4bbc-820e-ceac74e4d827', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '86bfab04-fd0a-4339-922e-241bb081e78e', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.386067] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Creating folder: Project (e1cd345e63584e108f502bb46ec3f02a). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 883.386240] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f126956-0d54-46aa-bb3e-07af727ec776 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.396833] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Created folder: Project (e1cd345e63584e108f502bb46ec3f02a) in parent group-v287607. [ 883.397031] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Creating folder: Instances. Parent ref: group-v287646. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 883.397271] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff835e88-cfaa-418c-bbae-c4357e767a4f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.406381] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Created folder: Instances in parent group-v287646. [ 883.406623] env[63355]: DEBUG oslo.service.loopingcall [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.406805] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 883.407012] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8601b695-fef7-4aa1-9383-883345c5fcd2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.425073] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.425073] env[63355]: value = "task-1349621" [ 883.425073] env[63355]: _type = "Task" [ 883.425073] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.434933] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349621, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.524855] env[63355]: DEBUG nova.compute.utils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.530095] env[63355]: DEBUG nova.compute.manager [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.530302] env[63355]: DEBUG nova.network.neutron [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 883.532186] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "5591bb55-83d7-4301-a3f9-fde945632344" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.532406] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "5591bb55-83d7-4301-a3f9-fde945632344" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.532601] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "5591bb55-83d7-4301-a3f9-fde945632344-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.532777] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "5591bb55-83d7-4301-a3f9-fde945632344-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.533027] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "5591bb55-83d7-4301-a3f9-fde945632344-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.535296] env[63355]: INFO nova.compute.manager [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Terminating instance [ 883.537374] env[63355]: DEBUG nova.compute.manager [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 883.537648] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 883.538405] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e5b62a-4449-4b4f-97e6-5223f3015267 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.546816] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.547152] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a11641d5-e85c-4998-bfdc-1f13ad39720a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.554795] env[63355]: DEBUG oslo_vmware.api [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 883.554795] env[63355]: value = "task-1349622" [ 883.554795] env[63355]: _type = "Task" [ 883.554795] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.565860] env[63355]: DEBUG oslo_vmware.api [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349622, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.574888] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.604545] env[63355]: DEBUG nova.policy [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4ec26bc0e3049b39a8e2c901e7b4d81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1cd345e63584e108f502bb46ec3f02a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.618589] env[63355]: DEBUG nova.compute.manager [req-aec77719-3273-40e9-8043-18805142b075 req-3391d41e-3e96-4245-8d36-9641d7e69e49 service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Received event network-changed-86bfab04-fd0a-4339-922e-241bb081e78e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.618747] env[63355]: DEBUG nova.compute.manager [req-aec77719-3273-40e9-8043-18805142b075 req-3391d41e-3e96-4245-8d36-9641d7e69e49 service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Refreshing instance network info cache due to event network-changed-86bfab04-fd0a-4339-922e-241bb081e78e. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 883.618902] env[63355]: DEBUG oslo_concurrency.lockutils [req-aec77719-3273-40e9-8043-18805142b075 req-3391d41e-3e96-4245-8d36-9641d7e69e49 service nova] Acquiring lock "refresh_cache-0b660647-697a-445b-bca4-82ef6975ea75" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.619458] env[63355]: DEBUG oslo_concurrency.lockutils [req-aec77719-3273-40e9-8043-18805142b075 req-3391d41e-3e96-4245-8d36-9641d7e69e49 service nova] Acquired lock "refresh_cache-0b660647-697a-445b-bca4-82ef6975ea75" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.619697] env[63355]: DEBUG nova.network.neutron [req-aec77719-3273-40e9-8043-18805142b075 req-3391d41e-3e96-4245-8d36-9641d7e69e49 service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Refreshing network info cache for port 86bfab04-fd0a-4339-922e-241bb081e78e {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.742581] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349618, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.850512] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c15384-6a5f-40fb-a4fc-5f7fd7c2a4b6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.858532] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa01013c-333e-416c-bb15-1d6777214a4d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.890237] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dccae7a-70eb-45b6-8164-5c2d5e5b0565 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.893463] env[63355]: DEBUG nova.network.neutron [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Successfully created port: ea0634b4-a639-4a82-9851-650958e3eff8 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.899903] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93feade-a9df-4dd3-8af6-982ea53e5133 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.915259] env[63355]: DEBUG nova.compute.provider_tree [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.934028] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349621, 'name': CreateVM_Task, 'duration_secs': 0.362436} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.934028] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 883.935030] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.935030] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.935152] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 883.935419] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ffc590f-9b0d-4aa0-9980-1a1afd955c17 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.939328] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 883.939328] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52758ebf-490e-1a8f-0b4d-238ef84703a3" [ 883.939328] env[63355]: _type = "Task" [ 883.939328] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.946523] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52758ebf-490e-1a8f-0b4d-238ef84703a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.030439] env[63355]: DEBUG nova.compute.manager [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.064579] env[63355]: DEBUG oslo_vmware.api [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349622, 'name': PowerOffVM_Task, 'duration_secs': 0.216362} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.064869] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.065050] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 884.065317] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-954fc5c3-b471-4576-abef-18e9a8f11bf8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.126466] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 884.126703] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 884.126882] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleting the datastore file [datastore2] 5591bb55-83d7-4301-a3f9-fde945632344 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.127501] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d2cbffa-2db1-4898-9591-46c3a9418c9a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.133128] env[63355]: DEBUG oslo_vmware.api [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 884.133128] env[63355]: value = "task-1349624" [ 884.133128] env[63355]: _type = "Task" [ 884.133128] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.141085] env[63355]: DEBUG oslo_vmware.api [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349624, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.239724] env[63355]: DEBUG oslo_vmware.api [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349618, 'name': PowerOnVM_Task, 'duration_secs': 0.536661} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.239983] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.240184] env[63355]: INFO nova.compute.manager [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Took 8.06 seconds to spawn the instance on the hypervisor. [ 884.240362] env[63355]: DEBUG nova.compute.manager [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.241146] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22932ac-c092-4a9c-8b29-157ee9f3fc10 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.421440] env[63355]: DEBUG nova.scheduler.client.report [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.450489] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52758ebf-490e-1a8f-0b4d-238ef84703a3, 'name': SearchDatastore_Task, 'duration_secs': 0.009287} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.450812] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.451053] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.451283] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.451441] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.451631] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.451888] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70b27867-6971-47ac-9726-6ddbc1ff9208 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.460962] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.461151] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 884.461880] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37211bb1-e3ce-4b77-8121-79ec23c247ed {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.467518] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 884.467518] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52887541-56a2-53d0-e1bd-54e22c962085" [ 884.467518] env[63355]: _type = "Task" [ 884.467518] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.474928] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52887541-56a2-53d0-e1bd-54e22c962085, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.608447] env[63355]: DEBUG nova.network.neutron [req-aec77719-3273-40e9-8043-18805142b075 req-3391d41e-3e96-4245-8d36-9641d7e69e49 service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Updated VIF entry in instance network info cache for port 86bfab04-fd0a-4339-922e-241bb081e78e. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 884.608898] env[63355]: DEBUG nova.network.neutron [req-aec77719-3273-40e9-8043-18805142b075 req-3391d41e-3e96-4245-8d36-9641d7e69e49 service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Updating instance_info_cache with network_info: [{"id": "86bfab04-fd0a-4339-922e-241bb081e78e", "address": "fa:16:3e:96:31:48", "network": {"id": "5930e4a2-0fe8-4b14-ad75-fca389ab1061", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.36", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfa2720fe39d48deb0cad174a7e5f3ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f39e3b37-7906-4bbc-820e-ceac74e4d827", "external-id": "nsx-vlan-transportzone-328", "segmentation_id": 328, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86bfab04-fd", "ovs_interfaceid": "86bfab04-fd0a-4339-922e-241bb081e78e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.643523] env[63355]: DEBUG oslo_vmware.api [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349624, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137976} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.644222] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.644722] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 884.645035] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 884.645435] env[63355]: INFO nova.compute.manager [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Took 1.11 seconds to destroy the instance on the hypervisor. [ 884.645805] env[63355]: DEBUG oslo.service.loopingcall [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.646105] env[63355]: DEBUG nova.compute.manager [-] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 884.646300] env[63355]: DEBUG nova.network.neutron [-] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 884.760627] env[63355]: INFO nova.compute.manager [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Took 41.06 seconds to build instance. [ 884.929936] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.910s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.932942] env[63355]: ERROR nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a83881a1-bb59-41b5-b08c-7999d67d97af, please check neutron logs for more information. [ 884.932942] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Traceback (most recent call last): [ 884.932942] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 884.932942] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] self.driver.spawn(context, instance, image_meta, [ 884.932942] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 884.932942] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] self._vmops.spawn(context, instance, image_meta, injected_files, [ 884.932942] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 884.932942] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] vm_ref = self.build_virtual_machine(instance, [ 884.932942] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 884.932942] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] vif_infos = vmwarevif.get_vif_info(self._session, [ 884.932942] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] for vif in network_info: [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] return self._sync_wrapper(fn, *args, **kwargs) [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] self.wait() [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] self[:] = self._gt.wait() [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] return self._exit_event.wait() [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] result = hub.switch() [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 884.933355] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] return self.greenlet.switch() [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] result = function(*args, **kwargs) [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] return func(*args, **kwargs) [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] raise e [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] nwinfo = self.network_api.allocate_for_instance( [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] created_port_ids = self._update_ports_for_instance( [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] with excutils.save_and_reraise_exception(): [ 884.933771] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] self.force_reraise() [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] raise self.value [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] updated_port = self._update_port( [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] _ensure_no_port_binding_failure(port) [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] raise exception.PortBindingFailed(port_id=port['id']) [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] nova.exception.PortBindingFailed: Binding failed for port a83881a1-bb59-41b5-b08c-7999d67d97af, please check neutron logs for more information. [ 884.934272] env[63355]: ERROR nova.compute.manager [instance: e49ba26c-486c-43aa-9595-231e7f31a522] [ 884.934649] env[63355]: DEBUG nova.compute.utils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Binding failed for port a83881a1-bb59-41b5-b08c-7999d67d97af, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 884.934974] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.762s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.939565] env[63355]: DEBUG nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Build of instance e49ba26c-486c-43aa-9595-231e7f31a522 was re-scheduled: Binding failed for port a83881a1-bb59-41b5-b08c-7999d67d97af, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 884.940017] env[63355]: DEBUG nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 884.940242] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "refresh_cache-e49ba26c-486c-43aa-9595-231e7f31a522" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.940389] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquired lock "refresh_cache-e49ba26c-486c-43aa-9595-231e7f31a522" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.940547] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.980103] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52887541-56a2-53d0-e1bd-54e22c962085, 'name': SearchDatastore_Task, 'duration_secs': 0.010776} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.980103] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98ceafeb-db37-43d7-801f-c20ac910c8f3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.985272] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 884.985272] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d59c23-4334-8bd8-48ea-6e019234f131" [ 884.985272] env[63355]: _type = "Task" [ 884.985272] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.000969] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d59c23-4334-8bd8-48ea-6e019234f131, 'name': SearchDatastore_Task, 'duration_secs': 0.009813} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.001324] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.003176] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 0b660647-697a-445b-bca4-82ef6975ea75/0b660647-697a-445b-bca4-82ef6975ea75.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 885.003176] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e79a225f-312d-44eb-9548-4b1427eff2c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.011096] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 885.011096] env[63355]: value = "task-1349625" [ 885.011096] env[63355]: _type = "Task" [ 885.011096] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.019998] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349625, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.043949] env[63355]: DEBUG nova.compute.manager [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 885.070889] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.071148] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.071305] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.071514] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.071666] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.071812] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.072042] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.072200] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.072367] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.072525] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.072693] env[63355]: DEBUG nova.virt.hardware [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.073562] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c969685e-dd60-438a-a086-0fd47d466ec3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.080988] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55541170-1cc8-4a06-932f-1421e021efd3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.111554] env[63355]: DEBUG oslo_concurrency.lockutils [req-aec77719-3273-40e9-8043-18805142b075 req-3391d41e-3e96-4245-8d36-9641d7e69e49 service nova] Releasing lock "refresh_cache-0b660647-697a-445b-bca4-82ef6975ea75" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.262482] env[63355]: DEBUG oslo_concurrency.lockutils [None req-db0ccd1e-22a8-4063-be9f-adcb9ea2d32f tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "386e847e-967b-4247-9730-cdc5ac251474" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 168.356s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.315110] env[63355]: DEBUG nova.compute.manager [req-859e8590-610f-486f-962c-3c66805be877 req-ca69355b-621e-4edf-a3fc-dfedcc0f0dfa service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Received event network-vif-plugged-ea0634b4-a639-4a82-9851-650958e3eff8 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.315279] env[63355]: DEBUG oslo_concurrency.lockutils [req-859e8590-610f-486f-962c-3c66805be877 req-ca69355b-621e-4edf-a3fc-dfedcc0f0dfa service nova] Acquiring lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.315729] env[63355]: DEBUG oslo_concurrency.lockutils [req-859e8590-610f-486f-962c-3c66805be877 req-ca69355b-621e-4edf-a3fc-dfedcc0f0dfa service nova] Lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.315985] env[63355]: DEBUG oslo_concurrency.lockutils [req-859e8590-610f-486f-962c-3c66805be877 req-ca69355b-621e-4edf-a3fc-dfedcc0f0dfa service nova] Lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.316465] env[63355]: DEBUG nova.compute.manager [req-859e8590-610f-486f-962c-3c66805be877 req-ca69355b-621e-4edf-a3fc-dfedcc0f0dfa service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] No waiting events found dispatching network-vif-plugged-ea0634b4-a639-4a82-9851-650958e3eff8 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 885.316890] env[63355]: WARNING nova.compute.manager [req-859e8590-610f-486f-962c-3c66805be877 req-ca69355b-621e-4edf-a3fc-dfedcc0f0dfa service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Received unexpected event network-vif-plugged-ea0634b4-a639-4a82-9851-650958e3eff8 for instance with vm_state building and task_state spawning. [ 885.418532] env[63355]: DEBUG nova.network.neutron [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Successfully updated port: ea0634b4-a639-4a82-9851-650958e3eff8 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 885.464047] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 885.510457] env[63355]: DEBUG nova.network.neutron [-] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.528407] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349625, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473809} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.531609] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 0b660647-697a-445b-bca4-82ef6975ea75/0b660647-697a-445b-bca4-82ef6975ea75.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 885.531999] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.532537] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9e084896-4f58-4b33-a9cd-a2ad86a0f2c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.539138] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 885.539138] env[63355]: value = "task-1349626" [ 885.539138] env[63355]: _type = "Task" [ 885.539138] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.546778] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349626, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.557869] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.646647] env[63355]: DEBUG nova.compute.manager [req-4e142218-75b5-4bd0-8abf-2b83d683646b req-3e9dbe26-9486-409d-8dd7-837474c5d58b service nova] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Received event network-vif-deleted-8fa3518a-34fb-4be4-992a-e08d9bfabe96 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.761544] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dd9dd7-5957-4956-ae2a-367fb30b3598 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.764830] env[63355]: DEBUG nova.compute.manager [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 885.774485] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588a8a5f-a754-4418-8d8a-ce541de5917c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.810683] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e870755-f862-4337-8977-5e0965ea692f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.818527] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d942d90-46c3-473b-b238-3aea1ca49888 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.832766] env[63355]: DEBUG nova.compute.provider_tree [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.921625] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "refresh_cache-b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.921775] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquired lock "refresh_cache-b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.921908] env[63355]: DEBUG nova.network.neutron [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.016525] env[63355]: INFO nova.compute.manager [-] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Took 1.37 seconds to deallocate network for instance. [ 886.049223] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349626, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056436} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.049497] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.050323] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd84fe3-94a8-4f33-a282-9e040a36942e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.069362] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Releasing lock "refresh_cache-e49ba26c-486c-43aa-9595-231e7f31a522" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.069631] env[63355]: DEBUG nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 886.069806] env[63355]: DEBUG nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 886.069966] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 886.080199] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 0b660647-697a-445b-bca4-82ef6975ea75/0b660647-697a-445b-bca4-82ef6975ea75.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.080867] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9fb72ce-e4f0-49ad-9d55-670b78474bd2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.101569] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 886.101569] env[63355]: value = "task-1349627" [ 886.101569] env[63355]: _type = "Task" [ 886.101569] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.111089] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349627, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.117051] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.289960] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.336337] env[63355]: DEBUG nova.scheduler.client.report [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.454007] env[63355]: DEBUG nova.network.neutron [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.523049] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.593963] env[63355]: DEBUG nova.network.neutron [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Updating instance_info_cache with network_info: [{"id": "ea0634b4-a639-4a82-9851-650958e3eff8", "address": "fa:16:3e:89:aa:a5", "network": {"id": "5930e4a2-0fe8-4b14-ad75-fca389ab1061", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfa2720fe39d48deb0cad174a7e5f3ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f39e3b37-7906-4bbc-820e-ceac74e4d827", "external-id": "nsx-vlan-transportzone-328", "segmentation_id": 328, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea0634b4-a6", "ovs_interfaceid": "ea0634b4-a639-4a82-9851-650958e3eff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.613408] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349627, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.619206] env[63355]: DEBUG nova.network.neutron [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.841351] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.906s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.841994] env[63355]: ERROR nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5a14c6da-f788-43f9-826a-4ce8b677fe35, please check neutron logs for more information. [ 886.841994] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Traceback (most recent call last): [ 886.841994] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 886.841994] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] self.driver.spawn(context, instance, image_meta, [ 886.841994] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 886.841994] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 886.841994] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 886.841994] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] vm_ref = self.build_virtual_machine(instance, [ 886.841994] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 886.841994] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] vif_infos = vmwarevif.get_vif_info(self._session, [ 886.841994] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] for vif in network_info: [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] return self._sync_wrapper(fn, *args, **kwargs) [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] self.wait() [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] self[:] = self._gt.wait() [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] return self._exit_event.wait() [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] current.throw(*self._exc) [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 886.842290] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] result = function(*args, **kwargs) [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] return func(*args, **kwargs) [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] raise e [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] nwinfo = self.network_api.allocate_for_instance( [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] created_port_ids = self._update_ports_for_instance( [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] with excutils.save_and_reraise_exception(): [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] self.force_reraise() [ 886.842609] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 886.842908] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] raise self.value [ 886.842908] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 886.842908] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] updated_port = self._update_port( [ 886.842908] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 886.842908] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] _ensure_no_port_binding_failure(port) [ 886.842908] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 886.842908] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] raise exception.PortBindingFailed(port_id=port['id']) [ 886.842908] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] nova.exception.PortBindingFailed: Binding failed for port 5a14c6da-f788-43f9-826a-4ce8b677fe35, please check neutron logs for more information. [ 886.842908] env[63355]: ERROR nova.compute.manager [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] [ 886.842908] env[63355]: DEBUG nova.compute.utils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Binding failed for port 5a14c6da-f788-43f9-826a-4ce8b677fe35, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 886.844024] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.328s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.846980] env[63355]: DEBUG nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Build of instance 7dd28cb7-d249-4b8e-beb3-c959f15e912f was re-scheduled: Binding failed for port 5a14c6da-f788-43f9-826a-4ce8b677fe35, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 886.848916] env[63355]: DEBUG nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 886.848916] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-7dd28cb7-d249-4b8e-beb3-c959f15e912f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.848916] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-7dd28cb7-d249-4b8e-beb3-c959f15e912f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.848916] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 887.096712] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Releasing lock "refresh_cache-b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.097058] env[63355]: DEBUG nova.compute.manager [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Instance network_info: |[{"id": "ea0634b4-a639-4a82-9851-650958e3eff8", "address": "fa:16:3e:89:aa:a5", "network": {"id": "5930e4a2-0fe8-4b14-ad75-fca389ab1061", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfa2720fe39d48deb0cad174a7e5f3ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f39e3b37-7906-4bbc-820e-ceac74e4d827", "external-id": "nsx-vlan-transportzone-328", "segmentation_id": 328, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea0634b4-a6", "ovs_interfaceid": "ea0634b4-a639-4a82-9851-650958e3eff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 887.097483] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:aa:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f39e3b37-7906-4bbc-820e-ceac74e4d827', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea0634b4-a639-4a82-9851-650958e3eff8', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.106065] env[63355]: DEBUG oslo.service.loopingcall [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.106403] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 887.109902] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa23b40a-c09d-4eec-8e94-acf48a0d45f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.124183] env[63355]: INFO nova.compute.manager [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: e49ba26c-486c-43aa-9595-231e7f31a522] Took 1.05 seconds to deallocate network for instance. [ 887.132168] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349627, 'name': ReconfigVM_Task, 'duration_secs': 0.670339} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.133415] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 0b660647-697a-445b-bca4-82ef6975ea75/0b660647-697a-445b-bca4-82ef6975ea75.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.134025] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.134025] env[63355]: value = "task-1349628" [ 887.134025] env[63355]: _type = "Task" [ 887.134025] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.134216] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b786c48c-0026-44cc-84ff-e11e423cee35 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.143919] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349628, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.145211] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 887.145211] env[63355]: value = "task-1349629" [ 887.145211] env[63355]: _type = "Task" [ 887.145211] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.155639] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349629, 'name': Rename_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.368124] env[63355]: DEBUG nova.compute.manager [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Received event network-changed-ea0634b4-a639-4a82-9851-650958e3eff8 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.368432] env[63355]: DEBUG nova.compute.manager [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Refreshing instance network info cache due to event network-changed-ea0634b4-a639-4a82-9851-650958e3eff8. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 887.368712] env[63355]: DEBUG oslo_concurrency.lockutils [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] Acquiring lock "refresh_cache-b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.369087] env[63355]: DEBUG oslo_concurrency.lockutils [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] Acquired lock "refresh_cache-b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.369457] env[63355]: DEBUG nova.network.neutron [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Refreshing network info cache for port ea0634b4-a639-4a82-9851-650958e3eff8 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 887.376078] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.461342] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.599542] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a92737-c354-4fdd-9012-ffe72c823112 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.608302] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad740ec1-1132-470b-a1e5-70c1890900da {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.643441] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfffa971-4f95-47ba-8aba-76778c331e89 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.651757] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349628, 'name': CreateVM_Task, 'duration_secs': 0.32489} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.653960] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.654672] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.654837] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.655171] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.656338] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd47fdcc-9d2d-4e49-bc57-82cfb484a6f3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.662588] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b036b50b-4116-4e40-abc2-f7213f6872ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.664010] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349629, 'name': Rename_Task, 'duration_secs': 0.154178} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.664541] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 887.665034] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8888999-344f-4da6-b5f1-d42070038e5f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.674682] env[63355]: DEBUG nova.compute.provider_tree [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.675858] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 887.675858] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52398aed-e1f0-ac44-6667-3987087b9ae8" [ 887.675858] env[63355]: _type = "Task" [ 887.675858] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.681508] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 887.681508] env[63355]: value = "task-1349630" [ 887.681508] env[63355]: _type = "Task" [ 887.681508] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.685293] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52398aed-e1f0-ac44-6667-3987087b9ae8, 'name': SearchDatastore_Task, 'duration_secs': 0.008181} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.688263] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.688482] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.688704] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.688845] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.689100] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.689555] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13074c6c-b9e8-4c3a-8e4c-9546a75c39ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.697928] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349630, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.698937] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.699117] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 887.699852] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bd32632-081b-4fab-b667-fc33d65b6c83 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.704387] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 887.704387] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52558712-4ca7-8f27-7e82-c3b0cec873a7" [ 887.704387] env[63355]: _type = "Task" [ 887.704387] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.712017] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52558712-4ca7-8f27-7e82-c3b0cec873a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.968036] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-7dd28cb7-d249-4b8e-beb3-c959f15e912f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.968036] env[63355]: DEBUG nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 887.968036] env[63355]: DEBUG nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 887.968036] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 887.981471] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 888.064741] env[63355]: DEBUG nova.network.neutron [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Updated VIF entry in instance network info cache for port ea0634b4-a639-4a82-9851-650958e3eff8. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 888.065221] env[63355]: DEBUG nova.network.neutron [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Updating instance_info_cache with network_info: [{"id": "ea0634b4-a639-4a82-9851-650958e3eff8", "address": "fa:16:3e:89:aa:a5", "network": {"id": "5930e4a2-0fe8-4b14-ad75-fca389ab1061", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfa2720fe39d48deb0cad174a7e5f3ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f39e3b37-7906-4bbc-820e-ceac74e4d827", "external-id": "nsx-vlan-transportzone-328", "segmentation_id": 328, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea0634b4-a6", "ovs_interfaceid": "ea0634b4-a639-4a82-9851-650958e3eff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.163348] env[63355]: INFO nova.scheduler.client.report [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Deleted allocations for instance e49ba26c-486c-43aa-9595-231e7f31a522 [ 888.178751] env[63355]: DEBUG nova.scheduler.client.report [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.194929] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349630, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.214682] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52558712-4ca7-8f27-7e82-c3b0cec873a7, 'name': SearchDatastore_Task, 'duration_secs': 0.009955} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.215577] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ae9e81b-e9ec-4ef5-801c-5c877526a63c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.220853] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 888.220853] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5298b380-7987-f8a6-7887-3d994a44d791" [ 888.220853] env[63355]: _type = "Task" [ 888.220853] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.229658] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5298b380-7987-f8a6-7887-3d994a44d791, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.483824] env[63355]: DEBUG nova.network.neutron [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.568226] env[63355]: DEBUG oslo_concurrency.lockutils [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] Releasing lock "refresh_cache-b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.568501] env[63355]: DEBUG nova.compute.manager [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received event network-changed-2348dcf0-b9a5-491f-87cb-a14e9e002368 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.568691] env[63355]: DEBUG nova.compute.manager [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Refreshing instance network info cache due to event network-changed-2348dcf0-b9a5-491f-87cb-a14e9e002368. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 888.568972] env[63355]: DEBUG oslo_concurrency.lockutils [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] Acquiring lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.569135] env[63355]: DEBUG oslo_concurrency.lockutils [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] Acquired lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.569298] env[63355]: DEBUG nova.network.neutron [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Refreshing network info cache for port 2348dcf0-b9a5-491f-87cb-a14e9e002368 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.670895] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4e6d13d-5b1a-44e2-ba3c-449b02fe6f01 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "e49ba26c-486c-43aa-9595-231e7f31a522" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 192.305s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.683012] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.839s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.683742] env[63355]: ERROR nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cb86c9bd-d267-4b95-a0ff-f75747ef143f, please check neutron logs for more information. [ 888.683742] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Traceback (most recent call last): [ 888.683742] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 888.683742] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] self.driver.spawn(context, instance, image_meta, [ 888.683742] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 888.683742] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 888.683742] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 888.683742] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] vm_ref = self.build_virtual_machine(instance, [ 888.683742] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 888.683742] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] vif_infos = vmwarevif.get_vif_info(self._session, [ 888.683742] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] for vif in network_info: [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] return self._sync_wrapper(fn, *args, **kwargs) [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] self.wait() [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] self[:] = self._gt.wait() [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] return self._exit_event.wait() [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] current.throw(*self._exc) [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 888.684205] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] result = function(*args, **kwargs) [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] return func(*args, **kwargs) [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] raise e [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] nwinfo = self.network_api.allocate_for_instance( [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] created_port_ids = self._update_ports_for_instance( [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] with excutils.save_and_reraise_exception(): [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] self.force_reraise() [ 888.684572] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 888.684993] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] raise self.value [ 888.684993] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 888.684993] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] updated_port = self._update_port( [ 888.684993] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 888.684993] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] _ensure_no_port_binding_failure(port) [ 888.684993] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 888.684993] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] raise exception.PortBindingFailed(port_id=port['id']) [ 888.684993] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] nova.exception.PortBindingFailed: Binding failed for port cb86c9bd-d267-4b95-a0ff-f75747ef143f, please check neutron logs for more information. [ 888.684993] env[63355]: ERROR nova.compute.manager [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] [ 888.684993] env[63355]: DEBUG nova.compute.utils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Binding failed for port cb86c9bd-d267-4b95-a0ff-f75747ef143f, please check neutron logs for more information. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 888.686298] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.795s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.687863] env[63355]: INFO nova.compute.claims [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.691242] env[63355]: DEBUG nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Build of instance 947a33d3-88c1-45ae-805c-eda2bc7626f1 was re-scheduled: Binding failed for port cb86c9bd-d267-4b95-a0ff-f75747ef143f, please check neutron logs for more information. {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 888.691852] env[63355]: DEBUG nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Unplugging VIFs for instance {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 888.692185] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-947a33d3-88c1-45ae-805c-eda2bc7626f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.692407] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-947a33d3-88c1-45ae-805c-eda2bc7626f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.692653] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.704453] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349630, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.731179] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5298b380-7987-f8a6-7887-3d994a44d791, 'name': SearchDatastore_Task, 'duration_secs': 0.010596} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.731444] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.731704] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda/b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 888.731955] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-48eecd6c-60bd-4f26-ac20-8b8d813727e6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.739476] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 888.739476] env[63355]: value = "task-1349631" [ 888.739476] env[63355]: _type = "Task" [ 888.739476] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.747299] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349631, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.987182] env[63355]: INFO nova.compute.manager [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 7dd28cb7-d249-4b8e-beb3-c959f15e912f] Took 1.02 seconds to deallocate network for instance. [ 889.174013] env[63355]: DEBUG nova.compute.manager [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.203764] env[63355]: DEBUG oslo_vmware.api [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349630, 'name': PowerOnVM_Task, 'duration_secs': 1.086051} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.204452] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.204557] env[63355]: INFO nova.compute.manager [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Took 8.53 seconds to spawn the instance on the hypervisor. [ 889.206217] env[63355]: DEBUG nova.compute.manager [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.206217] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c80786-58c4-4e01-877f-a643b5fbe61b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.220827] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.249548] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349631, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453933} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.249808] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda/b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 889.250025] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.250419] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f52a29e-c8fe-42f6-b9bd-5cf870743e2d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.257555] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 889.257555] env[63355]: value = "task-1349632" [ 889.257555] env[63355]: _type = "Task" [ 889.257555] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.266176] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349632, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.339373] env[63355]: DEBUG nova.network.neutron [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updated VIF entry in instance network info cache for port 2348dcf0-b9a5-491f-87cb-a14e9e002368. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.339774] env[63355]: DEBUG nova.network.neutron [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [{"id": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "address": "fa:16:3e:a9:13:18", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348dcf0-b9", "ovs_interfaceid": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.351246] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.695180] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.723036] env[63355]: INFO nova.compute.manager [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Took 31.26 seconds to build instance. [ 889.769604] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349632, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060805} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.769872] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 889.770669] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54422d5c-4db3-4a7e-94d7-2fc40e17ccaf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.794631] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda/b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.797184] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-430f32ac-5a99-408b-94ba-bc27542ae7f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.816069] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 889.816069] env[63355]: value = "task-1349633" [ 889.816069] env[63355]: _type = "Task" [ 889.816069] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.824795] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349633, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.842948] env[63355]: DEBUG oslo_concurrency.lockutils [req-fba80f96-298d-4854-8eae-f070a11efe24 req-749fca01-9737-4301-8699-6e45a947117d service nova] Releasing lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.853554] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-947a33d3-88c1-45ae-805c-eda2bc7626f1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.853892] env[63355]: DEBUG nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63355) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 889.854085] env[63355]: DEBUG nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.854580] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 889.874216] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 890.003580] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9dd96a8-a4f4-46e5-b227-eeb2700f8886 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.011460] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0529c6-c38b-40af-a3c8-49097d86de31 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.044552] env[63355]: INFO nova.scheduler.client.report [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted allocations for instance 7dd28cb7-d249-4b8e-beb3-c959f15e912f [ 890.050754] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6acb551-3b99-4aaa-9a2b-b4cdd2d10638 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.064520] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055e4e67-352a-44d6-8081-557019157d71 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.088596] env[63355]: DEBUG nova.compute.provider_tree [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.225596] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a7cdb215-60df-40cb-846c-19d0c11aca70 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "0b660647-697a-445b-bca4-82ef6975ea75" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 169.709s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.328360] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349633, 'name': ReconfigVM_Task, 'duration_secs': 0.268161} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.328360] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Reconfigured VM instance instance-0000003e to attach disk [datastore2] b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda/b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.328360] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7bd391d8-9b67-4973-bdc5-321f42ab4527 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.333618] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 890.333618] env[63355]: value = "task-1349634" [ 890.333618] env[63355]: _type = "Task" [ 890.333618] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.341357] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349634, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.376701] env[63355]: DEBUG nova.network.neutron [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.558035] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9b9f3cf5-622f-4493-a031-7455f9ab2ef5 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "7dd28cb7-d249-4b8e-beb3-c959f15e912f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 188.165s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.592558] env[63355]: DEBUG nova.scheduler.client.report [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.728441] env[63355]: DEBUG nova.compute.manager [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.848243] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349634, 'name': Rename_Task, 'duration_secs': 0.18282} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.848243] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 890.848243] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9b22442-aaf5-40e1-baa4-b778d89078c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.857199] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 890.857199] env[63355]: value = "task-1349635" [ 890.857199] env[63355]: _type = "Task" [ 890.857199] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.864191] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.879060] env[63355]: INFO nova.compute.manager [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 947a33d3-88c1-45ae-805c-eda2bc7626f1] Took 1.02 seconds to deallocate network for instance. [ 891.064389] env[63355]: DEBUG nova.compute.manager [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 891.104596] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.104884] env[63355]: DEBUG nova.compute.manager [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.108636] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.539s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.109472] env[63355]: INFO nova.compute.claims [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.257482] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.365036] env[63355]: DEBUG oslo_vmware.api [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349635, 'name': PowerOnVM_Task, 'duration_secs': 0.451164} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.365036] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 891.365205] env[63355]: INFO nova.compute.manager [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Took 6.32 seconds to spawn the instance on the hypervisor. [ 891.365324] env[63355]: DEBUG nova.compute.manager [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 891.366188] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf27875-5f8b-44f0-9179-0318e1a5da45 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.585315] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.614167] env[63355]: DEBUG nova.compute.utils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.617702] env[63355]: DEBUG nova.compute.manager [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.617923] env[63355]: DEBUG nova.network.neutron [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 891.704537] env[63355]: DEBUG nova.policy [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7123da3ee74b497395f1681e460aaaca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b6963b8e16b4986a4545914b75a38ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 891.892557] env[63355]: INFO nova.compute.manager [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Took 28.78 seconds to build instance. [ 891.916244] env[63355]: INFO nova.scheduler.client.report [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted allocations for instance 947a33d3-88c1-45ae-805c-eda2bc7626f1 [ 892.118412] env[63355]: DEBUG nova.compute.manager [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.214523] env[63355]: DEBUG nova.network.neutron [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Successfully created port: ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.390345] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1835f825-0d32-4029-9a3f-44a27c0106f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.394962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bcd9cf7c-66fd-4a47-a91d-8c609ece1b04 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 171.629s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.399026] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe1e45d-cfc7-4ac7-9dde-7bac21b316b8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.429508] env[63355]: DEBUG oslo_concurrency.lockutils [None req-87917f9f-de40-4d1f-a32f-1387a6da86af tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "947a33d3-88c1-45ae-805c-eda2bc7626f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 187.669s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.433805] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c09ecec-0b43-4541-a779-0ebcb63c1561 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.439982] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "6706c65c-3959-471b-82c0-2196f4aa1ab0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.440233] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "6706c65c-3959-471b-82c0-2196f4aa1ab0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.444654] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7c9933-d6bd-4c10-bdf9-2fea151e6777 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.458872] env[63355]: DEBUG nova.compute.provider_tree [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.611777] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.612459] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.899594] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 892.938334] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 892.961734] env[63355]: DEBUG nova.scheduler.client.report [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.130045] env[63355]: DEBUG nova.compute.manager [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.152577] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.152925] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.153065] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.153258] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.153410] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.153582] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.153790] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.153947] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.154132] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.154315] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.154506] env[63355]: DEBUG nova.virt.hardware [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.155376] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1686ef88-aa7b-46e7-9e3a-cdd8a405d016 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.166125] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35eb43e-de7e-4769-966f-66e5a90ef5cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.428714] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.464931] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.470416] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.362s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.471704] env[63355]: DEBUG nova.compute.manager [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.473821] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.638s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.475242] env[63355]: INFO nova.compute.claims [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.919705] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "1570cb36-76e0-4d06-8080-735b5246e92e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.919890] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "1570cb36-76e0-4d06-8080-735b5246e92e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.934513] env[63355]: DEBUG nova.compute.manager [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.935641] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9f0d26-172c-4fdc-8bfc-4a0680fec1ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.982565] env[63355]: DEBUG nova.compute.utils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.984448] env[63355]: DEBUG nova.compute.manager [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 893.984448] env[63355]: DEBUG nova.network.neutron [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 894.039211] env[63355]: DEBUG nova.compute.manager [req-1d9f7bed-7266-4265-805b-5df5f36b0e0e req-86dfb33a-1788-43c0-a6bd-021579500973 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Received event network-vif-plugged-ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.039528] env[63355]: DEBUG oslo_concurrency.lockutils [req-1d9f7bed-7266-4265-805b-5df5f36b0e0e req-86dfb33a-1788-43c0-a6bd-021579500973 service nova] Acquiring lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.039712] env[63355]: DEBUG oslo_concurrency.lockutils [req-1d9f7bed-7266-4265-805b-5df5f36b0e0e req-86dfb33a-1788-43c0-a6bd-021579500973 service nova] Lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.039932] env[63355]: DEBUG oslo_concurrency.lockutils [req-1d9f7bed-7266-4265-805b-5df5f36b0e0e req-86dfb33a-1788-43c0-a6bd-021579500973 service nova] Lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.040062] env[63355]: DEBUG nova.compute.manager [req-1d9f7bed-7266-4265-805b-5df5f36b0e0e req-86dfb33a-1788-43c0-a6bd-021579500973 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] No waiting events found dispatching network-vif-plugged-ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.041010] env[63355]: WARNING nova.compute.manager [req-1d9f7bed-7266-4265-805b-5df5f36b0e0e req-86dfb33a-1788-43c0-a6bd-021579500973 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Received unexpected event network-vif-plugged-ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5 for instance with vm_state building and task_state spawning. [ 894.061081] env[63355]: DEBUG nova.policy [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7bd1307924224936a7df31d9a3538d4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1efb1730acbf4ba5a19bd150bf99d48f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 894.131825] env[63355]: DEBUG nova.network.neutron [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Successfully updated port: ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.313769] env[63355]: DEBUG nova.network.neutron [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Successfully created port: b134b9dc-a12b-48b4-8051-39ec142ce649 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.447401] env[63355]: INFO nova.compute.manager [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] instance snapshotting [ 894.450429] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1571b3-0da7-4695-b6cd-1d646e8dee99 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.474850] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8413a3a3-ec00-488f-b558-687ae7ca1769 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.485448] env[63355]: DEBUG nova.compute.manager [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.637042] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "refresh_cache-f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.637192] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "refresh_cache-f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.637970] env[63355]: DEBUG nova.network.neutron [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.773717] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08f147a-e93c-4579-9dfd-0219f19bea8a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.781182] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2923e97-01ee-4ca5-b6c2-a2c5706d4584 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.810547] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e017fde-871c-49b3-83b8-ecf1aebd49b8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.817617] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d45f2de-59e4-44f7-b3ab-0e6d61a9efd5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.830397] env[63355]: DEBUG nova.compute.provider_tree [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.989110] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 894.989110] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ef350d09-213d-4719-91a6-21a5806b2899 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.995611] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 894.995611] env[63355]: value = "task-1349636" [ 894.995611] env[63355]: _type = "Task" [ 894.995611] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.006366] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349636, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.171167] env[63355]: DEBUG nova.network.neutron [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.324335] env[63355]: DEBUG nova.network.neutron [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Updating instance_info_cache with network_info: [{"id": "ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5", "address": "fa:16:3e:14:59:1a", "network": {"id": "60046874-e6ff-46c0-9843-752048f98e4a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1700834300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b6963b8e16b4986a4545914b75a38ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec8f312b-be", "ovs_interfaceid": "ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.332948] env[63355]: DEBUG nova.scheduler.client.report [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.497895] env[63355]: DEBUG nova.compute.manager [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.508617] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349636, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.520245] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.520574] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.520798] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.520991] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.521162] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.521307] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.521508] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.521661] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.521820] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.521976] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.522157] env[63355]: DEBUG nova.virt.hardware [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.522977] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd1e0a6-18e9-4ba0-83af-40ee4a3719ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.530652] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ca76cf-a293-42c6-af24-0929f07f2a7c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.730121] env[63355]: DEBUG nova.compute.manager [req-11c1b935-b1b1-4b05-a034-50fc10dd75a6 req-d1f61302-caa1-4cfa-8b75-2e471870ee80 service nova] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Received event network-vif-plugged-b134b9dc-a12b-48b4-8051-39ec142ce649 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.730343] env[63355]: DEBUG oslo_concurrency.lockutils [req-11c1b935-b1b1-4b05-a034-50fc10dd75a6 req-d1f61302-caa1-4cfa-8b75-2e471870ee80 service nova] Acquiring lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.730551] env[63355]: DEBUG oslo_concurrency.lockutils [req-11c1b935-b1b1-4b05-a034-50fc10dd75a6 req-d1f61302-caa1-4cfa-8b75-2e471870ee80 service nova] Lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.730718] env[63355]: DEBUG oslo_concurrency.lockutils [req-11c1b935-b1b1-4b05-a034-50fc10dd75a6 req-d1f61302-caa1-4cfa-8b75-2e471870ee80 service nova] Lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.730900] env[63355]: DEBUG nova.compute.manager [req-11c1b935-b1b1-4b05-a034-50fc10dd75a6 req-d1f61302-caa1-4cfa-8b75-2e471870ee80 service nova] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] No waiting events found dispatching network-vif-plugged-b134b9dc-a12b-48b4-8051-39ec142ce649 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.731198] env[63355]: WARNING nova.compute.manager [req-11c1b935-b1b1-4b05-a034-50fc10dd75a6 req-d1f61302-caa1-4cfa-8b75-2e471870ee80 service nova] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Received unexpected event network-vif-plugged-b134b9dc-a12b-48b4-8051-39ec142ce649 for instance with vm_state building and task_state spawning. [ 895.794281] env[63355]: DEBUG nova.network.neutron [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Successfully updated port: b134b9dc-a12b-48b4-8051-39ec142ce649 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.826233] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "refresh_cache-f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.826551] env[63355]: DEBUG nova.compute.manager [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Instance network_info: |[{"id": "ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5", "address": "fa:16:3e:14:59:1a", "network": {"id": "60046874-e6ff-46c0-9843-752048f98e4a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1700834300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b6963b8e16b4986a4545914b75a38ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec8f312b-be", "ovs_interfaceid": "ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.826967] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:59:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'df1bf911-aac9-4d2d-ae69-66ace3e6a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.834347] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Creating folder: Project (3b6963b8e16b4986a4545914b75a38ae). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 895.834605] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f325de5b-47f8-4d42-a064-9c829f55810b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.837591] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.838099] env[63355]: DEBUG nova.compute.manager [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 895.840519] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.502s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.840739] env[63355]: DEBUG nova.objects.instance [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lazy-loading 'resources' on Instance uuid 3d918cab-6ed6-4a37-a024-28e3db1b779c {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.846247] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Created folder: Project (3b6963b8e16b4986a4545914b75a38ae) in parent group-v287607. [ 895.846289] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Creating folder: Instances. Parent ref: group-v287651. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 895.846507] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd81ee19-8578-478e-870b-44c8a5d00da8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.856109] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Created folder: Instances in parent group-v287651. [ 895.856766] env[63355]: DEBUG oslo.service.loopingcall [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.856766] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 895.856897] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28c0ad23-b3d1-4819-afac-a1bcd9a86878 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.876723] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.876723] env[63355]: value = "task-1349639" [ 895.876723] env[63355]: _type = "Task" [ 895.876723] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.884927] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349639, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.009810] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349636, 'name': CreateSnapshot_Task, 'duration_secs': 0.831077} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.010107] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 896.010936] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf20316-9c49-4b8a-88cc-5ee98958a7ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.070314] env[63355]: DEBUG nova.compute.manager [req-f327289b-f912-4266-87dc-cae7f365b0a1 req-00688c3d-3af9-46e0-82fe-3a02f9be55f9 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Received event network-changed-ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.071960] env[63355]: DEBUG nova.compute.manager [req-f327289b-f912-4266-87dc-cae7f365b0a1 req-00688c3d-3af9-46e0-82fe-3a02f9be55f9 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Refreshing instance network info cache due to event network-changed-ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.071960] env[63355]: DEBUG oslo_concurrency.lockutils [req-f327289b-f912-4266-87dc-cae7f365b0a1 req-00688c3d-3af9-46e0-82fe-3a02f9be55f9 service nova] Acquiring lock "refresh_cache-f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.071960] env[63355]: DEBUG oslo_concurrency.lockutils [req-f327289b-f912-4266-87dc-cae7f365b0a1 req-00688c3d-3af9-46e0-82fe-3a02f9be55f9 service nova] Acquired lock "refresh_cache-f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.071960] env[63355]: DEBUG nova.network.neutron [req-f327289b-f912-4266-87dc-cae7f365b0a1 req-00688c3d-3af9-46e0-82fe-3a02f9be55f9 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Refreshing network info cache for port ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.298027] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "refresh_cache-8e3ec9d3-bc22-4e39-ad7c-93268dd59020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.298027] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "refresh_cache-8e3ec9d3-bc22-4e39-ad7c-93268dd59020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.298027] env[63355]: DEBUG nova.network.neutron [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.344400] env[63355]: DEBUG nova.compute.utils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.348427] env[63355]: DEBUG nova.compute.manager [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 896.348600] env[63355]: DEBUG nova.network.neutron [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.386071] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349639, 'name': CreateVM_Task, 'duration_secs': 0.330922} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.388421] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 896.389464] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.389464] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.389811] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.390270] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0fe6cf6-ad6b-4edc-969a-6cf75d844349 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.392897] env[63355]: DEBUG nova.policy [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0f5e88f84d54fbe8f418fcb71a841a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4a7aa4a2115c4aedbca273eaf0240864', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.397574] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 896.397574] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526da91f-c730-b97b-1974-0cc6fb4be97a" [ 896.397574] env[63355]: _type = "Task" [ 896.397574] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.408102] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526da91f-c730-b97b-1974-0cc6fb4be97a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.528615] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 896.529359] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-89dfef51-6bcd-41e8-9700-795390307609 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.537564] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 896.537564] env[63355]: value = "task-1349640" [ 896.537564] env[63355]: _type = "Task" [ 896.537564] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.548586] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349640, 'name': CloneVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.658212] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f802c6b0-2669-42af-958b-1b4738daee0b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.662447] env[63355]: DEBUG nova.network.neutron [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Successfully created port: fd4a5737-d1df-4e4e-aa70-3002a5778567 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 896.668966] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba1a039-d6b0-4ab1-9533-e6f5c1e19da2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.706093] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4396a5be-edec-48d1-a759-009b983e6a11 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.713701] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42917e62-6e49-433d-a0ba-ad6c761b2dd3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.728602] env[63355]: DEBUG nova.compute.provider_tree [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.841112] env[63355]: DEBUG nova.network.neutron [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 896.848977] env[63355]: DEBUG nova.compute.manager [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 896.911137] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526da91f-c730-b97b-1974-0cc6fb4be97a, 'name': SearchDatastore_Task, 'duration_secs': 0.010606} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.911432] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.911673] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.911900] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.912112] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.912891] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.912891] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b598b0ac-c26b-4aeb-8766-5125679cc806 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.922474] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.922656] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 896.924182] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ef16b32-04ed-45a3-8c3d-4013cfdd924e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.929446] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 896.929446] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523f5912-0370-271e-0e06-48876e59414f" [ 896.929446] env[63355]: _type = "Task" [ 896.929446] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.937193] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523f5912-0370-271e-0e06-48876e59414f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.992870] env[63355]: DEBUG nova.network.neutron [req-f327289b-f912-4266-87dc-cae7f365b0a1 req-00688c3d-3af9-46e0-82fe-3a02f9be55f9 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Updated VIF entry in instance network info cache for port ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 896.992870] env[63355]: DEBUG nova.network.neutron [req-f327289b-f912-4266-87dc-cae7f365b0a1 req-00688c3d-3af9-46e0-82fe-3a02f9be55f9 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Updating instance_info_cache with network_info: [{"id": "ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5", "address": "fa:16:3e:14:59:1a", "network": {"id": "60046874-e6ff-46c0-9843-752048f98e4a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1700834300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b6963b8e16b4986a4545914b75a38ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec8f312b-be", "ovs_interfaceid": "ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.038831] env[63355]: DEBUG nova.network.neutron [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Updating instance_info_cache with network_info: [{"id": "b134b9dc-a12b-48b4-8051-39ec142ce649", "address": "fa:16:3e:30:2c:29", "network": {"id": "26b0dabd-e662-4b47-bcd4-a50e05f0a52d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-825312565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1efb1730acbf4ba5a19bd150bf99d48f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb134b9dc-a1", "ovs_interfaceid": "b134b9dc-a12b-48b4-8051-39ec142ce649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.049680] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349640, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.234293] env[63355]: DEBUG nova.scheduler.client.report [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.439380] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523f5912-0370-271e-0e06-48876e59414f, 'name': SearchDatastore_Task, 'duration_secs': 0.010203} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.440155] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37d1b824-bf53-43e1-8b93-d068269c024b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.445742] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 897.445742] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]520ddbdf-58a9-62a0-16d1-a6696a5439d7" [ 897.445742] env[63355]: _type = "Task" [ 897.445742] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.452959] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]520ddbdf-58a9-62a0-16d1-a6696a5439d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.497125] env[63355]: DEBUG oslo_concurrency.lockutils [req-f327289b-f912-4266-87dc-cae7f365b0a1 req-00688c3d-3af9-46e0-82fe-3a02f9be55f9 service nova] Releasing lock "refresh_cache-f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.546047] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "refresh_cache-8e3ec9d3-bc22-4e39-ad7c-93268dd59020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.546047] env[63355]: DEBUG nova.compute.manager [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Instance network_info: |[{"id": "b134b9dc-a12b-48b4-8051-39ec142ce649", "address": "fa:16:3e:30:2c:29", "network": {"id": "26b0dabd-e662-4b47-bcd4-a50e05f0a52d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-825312565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1efb1730acbf4ba5a19bd150bf99d48f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb134b9dc-a1", "ovs_interfaceid": "b134b9dc-a12b-48b4-8051-39ec142ce649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.546262] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:2c:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd5970ab5-34b8-4065-bfa6-f568b8f103b7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b134b9dc-a12b-48b4-8051-39ec142ce649', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.553532] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Creating folder: Project (1efb1730acbf4ba5a19bd150bf99d48f). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 897.554343] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4bfaded4-a5d0-4a0b-bf07-8bf050cbbeba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.559832] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349640, 'name': CloneVM_Task, 'duration_secs': 0.970032} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.560154] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Created linked-clone VM from snapshot [ 897.560863] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2244e3c-b066-460f-b934-96eb94b06239 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.564832] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Created folder: Project (1efb1730acbf4ba5a19bd150bf99d48f) in parent group-v287607. [ 897.565042] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Creating folder: Instances. Parent ref: group-v287655. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 897.567931] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-26dcab28-d1ea-454d-a5c2-136b62946eb8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.569366] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Uploading image 2584bf77-1454-4062-a175-2abf8f10d255 {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 897.579643] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Created folder: Instances in parent group-v287655. [ 897.579643] env[63355]: DEBUG oslo.service.loopingcall [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.580089] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.580089] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91090b31-3b52-4014-a325-7b92d8a82bd1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.600870] env[63355]: DEBUG oslo_vmware.rw_handles [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 897.600870] env[63355]: value = "vm-287654" [ 897.600870] env[63355]: _type = "VirtualMachine" [ 897.600870] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 897.601125] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5c442329-34c8-4c5b-8cbd-22399e75dc6f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.603321] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.603321] env[63355]: value = "task-1349643" [ 897.603321] env[63355]: _type = "Task" [ 897.603321] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.607486] env[63355]: DEBUG oslo_vmware.rw_handles [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lease: (returnval){ [ 897.607486] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c6a73f-2757-bdf4-7ee4-9357852b448f" [ 897.607486] env[63355]: _type = "HttpNfcLease" [ 897.607486] env[63355]: } obtained for exporting VM: (result){ [ 897.607486] env[63355]: value = "vm-287654" [ 897.607486] env[63355]: _type = "VirtualMachine" [ 897.607486] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 897.607697] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the lease: (returnval){ [ 897.607697] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c6a73f-2757-bdf4-7ee4-9357852b448f" [ 897.607697] env[63355]: _type = "HttpNfcLease" [ 897.607697] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 897.613415] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349643, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.616326] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 897.616326] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c6a73f-2757-bdf4-7ee4-9357852b448f" [ 897.616326] env[63355]: _type = "HttpNfcLease" [ 897.616326] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 897.740242] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.899s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.743848] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.169s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.746040] env[63355]: INFO nova.compute.claims [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 897.771339] env[63355]: INFO nova.scheduler.client.report [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Deleted allocations for instance 3d918cab-6ed6-4a37-a024-28e3db1b779c [ 897.780281] env[63355]: DEBUG nova.compute.manager [req-ad1e1b50-80ab-49ab-a40d-45cc75e22abc req-30d68849-d4a1-4b64-afd5-3926d9a2a55a service nova] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Received event network-changed-b134b9dc-a12b-48b4-8051-39ec142ce649 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.780523] env[63355]: DEBUG nova.compute.manager [req-ad1e1b50-80ab-49ab-a40d-45cc75e22abc req-30d68849-d4a1-4b64-afd5-3926d9a2a55a service nova] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Refreshing instance network info cache due to event network-changed-b134b9dc-a12b-48b4-8051-39ec142ce649. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 897.780738] env[63355]: DEBUG oslo_concurrency.lockutils [req-ad1e1b50-80ab-49ab-a40d-45cc75e22abc req-30d68849-d4a1-4b64-afd5-3926d9a2a55a service nova] Acquiring lock "refresh_cache-8e3ec9d3-bc22-4e39-ad7c-93268dd59020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.780914] env[63355]: DEBUG oslo_concurrency.lockutils [req-ad1e1b50-80ab-49ab-a40d-45cc75e22abc req-30d68849-d4a1-4b64-afd5-3926d9a2a55a service nova] Acquired lock "refresh_cache-8e3ec9d3-bc22-4e39-ad7c-93268dd59020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.781129] env[63355]: DEBUG nova.network.neutron [req-ad1e1b50-80ab-49ab-a40d-45cc75e22abc req-30d68849-d4a1-4b64-afd5-3926d9a2a55a service nova] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Refreshing network info cache for port b134b9dc-a12b-48b4-8051-39ec142ce649 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.862080] env[63355]: DEBUG nova.compute.manager [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 897.883422] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.883671] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.883830] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.884017] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.884177] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.884323] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.884531] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.884689] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.884918] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.885131] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.885311] env[63355]: DEBUG nova.virt.hardware [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.886422] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5018a4-a650-4030-b5ef-178fe7c2df15 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.895460] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b1986d-9d0b-4627-93c9-00e9860dda55 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.955657] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]520ddbdf-58a9-62a0-16d1-a6696a5439d7, 'name': SearchDatastore_Task, 'duration_secs': 0.031705} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.955993] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.956280] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] f5e62ce1-40b7-4648-a4a6-068ff06eaf9b/f5e62ce1-40b7-4648-a4a6-068ff06eaf9b.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 897.956552] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-176ad76c-a620-42c4-91d8-87feb2407490 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.963313] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 897.963313] env[63355]: value = "task-1349645" [ 897.963313] env[63355]: _type = "Task" [ 897.963313] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.971203] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349645, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.117673] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349643, 'name': CreateVM_Task, 'duration_secs': 0.408053} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.119362] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 898.119604] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 898.119604] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c6a73f-2757-bdf4-7ee4-9357852b448f" [ 898.119604] env[63355]: _type = "HttpNfcLease" [ 898.119604] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 898.120268] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.120432] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.120740] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.121044] env[63355]: DEBUG oslo_vmware.rw_handles [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 898.121044] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c6a73f-2757-bdf4-7ee4-9357852b448f" [ 898.121044] env[63355]: _type = "HttpNfcLease" [ 898.121044] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 898.121262] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c937e212-6118-4044-9a41-a73255b6b85a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.123272] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed100cbc-11e4-4cb0-8438-11045e0e28c5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.133812] env[63355]: DEBUG oslo_vmware.rw_handles [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529066c3-c3ca-f757-fb1d-9dc2d1b34bb0/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 898.134085] env[63355]: DEBUG oslo_vmware.rw_handles [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529066c3-c3ca-f757-fb1d-9dc2d1b34bb0/disk-0.vmdk for reading. {{(pid=63355) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 898.135661] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 898.135661] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f4f7f6-2827-e5ca-f9ba-7f4a5f511f16" [ 898.135661] env[63355]: _type = "Task" [ 898.135661] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.201819] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f4f7f6-2827-e5ca-f9ba-7f4a5f511f16, 'name': SearchDatastore_Task, 'duration_secs': 0.011058} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.203458] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.203867] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.204219] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.204423] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.204668] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.205333] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f09a9d98-28a2-441f-a20d-9af5310b2b85 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.217365] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.217610] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.218440] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bb3ff4f-2bee-42b6-965f-c08c187f058c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.224487] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 898.224487] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52eb1eb2-a0bf-08cc-cbd0-51e61e84b055" [ 898.224487] env[63355]: _type = "Task" [ 898.224487] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.233378] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52eb1eb2-a0bf-08cc-cbd0-51e61e84b055, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.236244] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-912922e4-c2da-4186-aba3-42997e62c600 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.288324] env[63355]: DEBUG oslo_concurrency.lockutils [None req-75872444-9d17-4fc0-b68b-3f02c93dec7f tempest-ServerTagsTestJSON-797259541 tempest-ServerTagsTestJSON-797259541-project-member] Lock "3d918cab-6ed6-4a37-a024-28e3db1b779c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.987s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.478654] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349645, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.663400] env[63355]: DEBUG nova.network.neutron [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Successfully updated port: fd4a5737-d1df-4e4e-aa70-3002a5778567 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.744331] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52eb1eb2-a0bf-08cc-cbd0-51e61e84b055, 'name': SearchDatastore_Task, 'duration_secs': 0.00917} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.745899] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdc23c26-41f3-48d2-90ec-36d525dd4965 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.753267] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 898.753267] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5231b0ad-32b7-9670-90f7-fdd35f922d9d" [ 898.753267] env[63355]: _type = "Task" [ 898.753267] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.766168] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5231b0ad-32b7-9670-90f7-fdd35f922d9d, 'name': SearchDatastore_Task, 'duration_secs': 0.009378} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.767152] env[63355]: DEBUG nova.network.neutron [req-ad1e1b50-80ab-49ab-a40d-45cc75e22abc req-30d68849-d4a1-4b64-afd5-3926d9a2a55a service nova] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Updated VIF entry in instance network info cache for port b134b9dc-a12b-48b4-8051-39ec142ce649. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 898.767801] env[63355]: DEBUG nova.network.neutron [req-ad1e1b50-80ab-49ab-a40d-45cc75e22abc req-30d68849-d4a1-4b64-afd5-3926d9a2a55a service nova] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Updating instance_info_cache with network_info: [{"id": "b134b9dc-a12b-48b4-8051-39ec142ce649", "address": "fa:16:3e:30:2c:29", "network": {"id": "26b0dabd-e662-4b47-bcd4-a50e05f0a52d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-825312565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1efb1730acbf4ba5a19bd150bf99d48f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb134b9dc-a1", "ovs_interfaceid": "b134b9dc-a12b-48b4-8051-39ec142ce649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.769369] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.769840] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 8e3ec9d3-bc22-4e39-ad7c-93268dd59020/8e3ec9d3-bc22-4e39-ad7c-93268dd59020.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 898.770739] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-acb71663-8728-429e-bf1d-7ddb001c2e77 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.780148] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 898.780148] env[63355]: value = "task-1349646" [ 898.780148] env[63355]: _type = "Task" [ 898.780148] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.793397] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349646, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.985711] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349645, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677529} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.991772] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] f5e62ce1-40b7-4648-a4a6-068ff06eaf9b/f5e62ce1-40b7-4648-a4a6-068ff06eaf9b.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.991772] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.991772] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2041929-dd6c-4b47-97c8-7d5accff0247 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.997239] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 898.997239] env[63355]: value = "task-1349647" [ 898.997239] env[63355]: _type = "Task" [ 898.997239] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.007590] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349647, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.123932] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0ebabf-501e-4b8a-9865-ed078a329e8a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.133804] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4331a3-8a6a-4a5d-9958-cb793878c335 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.174277] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquiring lock "refresh_cache-b6056441-9ee4-484f-a1d2-077546f2c581" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.174475] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquired lock "refresh_cache-b6056441-9ee4-484f-a1d2-077546f2c581" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.174846] env[63355]: DEBUG nova.network.neutron [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.177561] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac857255-31cb-4a8a-971b-610f831fd182 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.187180] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c72490e-b9f4-4162-9956-b011fa674317 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.205740] env[63355]: DEBUG nova.compute.provider_tree [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.273493] env[63355]: DEBUG oslo_concurrency.lockutils [req-ad1e1b50-80ab-49ab-a40d-45cc75e22abc req-30d68849-d4a1-4b64-afd5-3926d9a2a55a service nova] Releasing lock "refresh_cache-8e3ec9d3-bc22-4e39-ad7c-93268dd59020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.292046] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349646, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.508501] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349647, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070617} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.508787] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.509603] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed749de5-58b8-4144-bbb1-5403379e970d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.534759] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] f5e62ce1-40b7-4648-a4a6-068ff06eaf9b/f5e62ce1-40b7-4648-a4a6-068ff06eaf9b.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.534759] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-beb28b5e-acb4-4b6c-bd38-58c5c2a3545b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.557590] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 899.557590] env[63355]: value = "task-1349648" [ 899.557590] env[63355]: _type = "Task" [ 899.557590] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.567712] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349648, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.710780] env[63355]: DEBUG nova.scheduler.client.report [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.714914] env[63355]: DEBUG nova.network.neutron [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 899.790465] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349646, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681553} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.790899] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 8e3ec9d3-bc22-4e39-ad7c-93268dd59020/8e3ec9d3-bc22-4e39-ad7c-93268dd59020.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 899.791319] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 899.791519] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7bdf8754-a0b5-45fd-a3bb-2d26911596b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.798893] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 899.798893] env[63355]: value = "task-1349649" [ 899.798893] env[63355]: _type = "Task" [ 899.798893] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.807183] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349649, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.810020] env[63355]: DEBUG nova.compute.manager [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Received event network-vif-plugged-fd4a5737-d1df-4e4e-aa70-3002a5778567 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.810395] env[63355]: DEBUG oslo_concurrency.lockutils [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] Acquiring lock "b6056441-9ee4-484f-a1d2-077546f2c581-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.810670] env[63355]: DEBUG oslo_concurrency.lockutils [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] Lock "b6056441-9ee4-484f-a1d2-077546f2c581-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.810889] env[63355]: DEBUG oslo_concurrency.lockutils [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] Lock "b6056441-9ee4-484f-a1d2-077546f2c581-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.811129] env[63355]: DEBUG nova.compute.manager [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] No waiting events found dispatching network-vif-plugged-fd4a5737-d1df-4e4e-aa70-3002a5778567 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.811349] env[63355]: WARNING nova.compute.manager [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Received unexpected event network-vif-plugged-fd4a5737-d1df-4e4e-aa70-3002a5778567 for instance with vm_state building and task_state spawning. [ 899.811607] env[63355]: DEBUG nova.compute.manager [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Received event network-changed-fd4a5737-d1df-4e4e-aa70-3002a5778567 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.812079] env[63355]: DEBUG nova.compute.manager [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Refreshing instance network info cache due to event network-changed-fd4a5737-d1df-4e4e-aa70-3002a5778567. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 899.812549] env[63355]: DEBUG oslo_concurrency.lockutils [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] Acquiring lock "refresh_cache-b6056441-9ee4-484f-a1d2-077546f2c581" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.877455] env[63355]: DEBUG nova.network.neutron [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Updating instance_info_cache with network_info: [{"id": "fd4a5737-d1df-4e4e-aa70-3002a5778567", "address": "fa:16:3e:ae:f8:52", "network": {"id": "99e43a47-c612-4732-ade1-b697d26e2054", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1433466117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a7aa4a2115c4aedbca273eaf0240864", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd4a5737-d1", "ovs_interfaceid": "fd4a5737-d1df-4e4e-aa70-3002a5778567", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.070424] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349648, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.218756] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.219455] env[63355]: DEBUG nova.compute.manager [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 900.222158] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.932s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.223709] env[63355]: INFO nova.compute.claims [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.309904] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349649, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068746} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.310787] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 900.311477] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7f21fe-6d2a-4119-9d2f-c36841a77a83 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.338135] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 8e3ec9d3-bc22-4e39-ad7c-93268dd59020/8e3ec9d3-bc22-4e39-ad7c-93268dd59020.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.339101] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b05b1edc-fd00-43ec-a8d0-bdd79dc83b2f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.361815] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 900.361815] env[63355]: value = "task-1349650" [ 900.361815] env[63355]: _type = "Task" [ 900.361815] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.368929] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349650, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.380913] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Releasing lock "refresh_cache-b6056441-9ee4-484f-a1d2-077546f2c581" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.380996] env[63355]: DEBUG nova.compute.manager [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Instance network_info: |[{"id": "fd4a5737-d1df-4e4e-aa70-3002a5778567", "address": "fa:16:3e:ae:f8:52", "network": {"id": "99e43a47-c612-4732-ade1-b697d26e2054", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1433466117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a7aa4a2115c4aedbca273eaf0240864", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd4a5737-d1", "ovs_interfaceid": "fd4a5737-d1df-4e4e-aa70-3002a5778567", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 900.381756] env[63355]: DEBUG oslo_concurrency.lockutils [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] Acquired lock "refresh_cache-b6056441-9ee4-484f-a1d2-077546f2c581" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.381756] env[63355]: DEBUG nova.network.neutron [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Refreshing network info cache for port fd4a5737-d1df-4e4e-aa70-3002a5778567 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.382887] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:f8:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccf76700-491b-4462-ab19-e6d3a9ff87ac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fd4a5737-d1df-4e4e-aa70-3002a5778567', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.391010] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Creating folder: Project (4a7aa4a2115c4aedbca273eaf0240864). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 900.392074] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc3349e8-b99b-46b4-8275-486253fbcd05 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.403697] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Created folder: Project (4a7aa4a2115c4aedbca273eaf0240864) in parent group-v287607. [ 900.403952] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Creating folder: Instances. Parent ref: group-v287658. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 900.404351] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2c2ad3b4-ab0b-4953-80e4-4d971dd93595 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.413710] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Created folder: Instances in parent group-v287658. [ 900.413921] env[63355]: DEBUG oslo.service.loopingcall [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.414127] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.414338] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-593bf10f-02de-4ff0-8f8e-03dbe4fcf530 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.434270] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.434270] env[63355]: value = "task-1349653" [ 900.434270] env[63355]: _type = "Task" [ 900.434270] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.442044] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349653, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.568063] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349648, 'name': ReconfigVM_Task, 'duration_secs': 0.526043} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.568314] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Reconfigured VM instance instance-0000003f to attach disk [datastore2] f5e62ce1-40b7-4648-a4a6-068ff06eaf9b/f5e62ce1-40b7-4648-a4a6-068ff06eaf9b.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.568957] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4855cf6d-7b65-4bf7-a2bf-91a7871f8379 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.576172] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 900.576172] env[63355]: value = "task-1349654" [ 900.576172] env[63355]: _type = "Task" [ 900.576172] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.584523] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349654, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.728517] env[63355]: DEBUG nova.compute.utils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 900.731903] env[63355]: DEBUG nova.compute.manager [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Not allocating networking since 'none' was specified. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 900.870494] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349650, 'name': ReconfigVM_Task, 'duration_secs': 0.416429} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.870747] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 8e3ec9d3-bc22-4e39-ad7c-93268dd59020/8e3ec9d3-bc22-4e39-ad7c-93268dd59020.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.871485] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-23886816-3808-4762-bd04-115260776e8e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.878099] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 900.878099] env[63355]: value = "task-1349655" [ 900.878099] env[63355]: _type = "Task" [ 900.878099] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.887513] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349655, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.950591] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349653, 'name': CreateVM_Task, 'duration_secs': 0.386241} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.950729] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 900.951642] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.951712] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.952708] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.953024] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd3cfc7b-6b5e-4d3f-aba9-1fe1a96f2246 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.958377] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for the task: (returnval){ [ 900.958377] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522ffe48-a113-c17f-c0ba-590740f614f4" [ 900.958377] env[63355]: _type = "Task" [ 900.958377] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.968072] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522ffe48-a113-c17f-c0ba-590740f614f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.086281] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349654, 'name': Rename_Task, 'duration_secs': 0.217005} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.086491] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 901.086672] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f64ce3e9-aa72-4918-9330-684bc0bdfd2f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.092386] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 901.092386] env[63355]: value = "task-1349656" [ 901.092386] env[63355]: _type = "Task" [ 901.092386] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.102719] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349656, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.199692] env[63355]: DEBUG nova.network.neutron [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Updated VIF entry in instance network info cache for port fd4a5737-d1df-4e4e-aa70-3002a5778567. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.200080] env[63355]: DEBUG nova.network.neutron [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Updating instance_info_cache with network_info: [{"id": "fd4a5737-d1df-4e4e-aa70-3002a5778567", "address": "fa:16:3e:ae:f8:52", "network": {"id": "99e43a47-c612-4732-ade1-b697d26e2054", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1433466117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a7aa4a2115c4aedbca273eaf0240864", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd4a5737-d1", "ovs_interfaceid": "fd4a5737-d1df-4e4e-aa70-3002a5778567", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.236283] env[63355]: DEBUG nova.compute.manager [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 901.389992] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349655, 'name': Rename_Task, 'duration_secs': 0.16805} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.392494] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 901.392926] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-030551da-0a39-47ef-9fee-caa24a7fa76d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.399744] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 901.399744] env[63355]: value = "task-1349657" [ 901.399744] env[63355]: _type = "Task" [ 901.399744] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.411100] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349657, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.471202] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522ffe48-a113-c17f-c0ba-590740f614f4, 'name': SearchDatastore_Task, 'duration_secs': 0.013357} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.471531] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.471767] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.472013] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.472173] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.472358] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.472621] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e87569ff-6ef0-4ebc-a452-fac78bd01c8c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.481030] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.481030] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.481759] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f946e9cd-b1db-4d4c-89af-9a4d5889836f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.486733] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for the task: (returnval){ [ 901.486733] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526aa95d-42d0-1022-c8d2-4e042d2892ee" [ 901.486733] env[63355]: _type = "Task" [ 901.486733] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.497108] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526aa95d-42d0-1022-c8d2-4e042d2892ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.515994] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1534c3e-4557-4367-b7b1-ec7d97df8103 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.523220] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710b2478-4c89-4e3d-a5df-00ab63f546ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.554974] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a4dd69-5c71-470d-a221-45207224bd81 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.562692] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fde82e-77d8-405f-bf0a-fb8c2acc239c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.577448] env[63355]: DEBUG nova.compute.provider_tree [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.603348] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349656, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.702921] env[63355]: DEBUG oslo_concurrency.lockutils [req-0c24a4bf-7740-43dd-8cee-029861c55860 req-63651ab6-e9f2-482a-aec1-3a7988bd2c27 service nova] Releasing lock "refresh_cache-b6056441-9ee4-484f-a1d2-077546f2c581" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.910644] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349657, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.998021] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526aa95d-42d0-1022-c8d2-4e042d2892ee, 'name': SearchDatastore_Task, 'duration_secs': 0.01151} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.998896] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3abc2456-d493-4eb3-817b-07d8c6fa0d70 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.004480] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for the task: (returnval){ [ 902.004480] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525592c4-8eb8-c516-8292-b8972746863d" [ 902.004480] env[63355]: _type = "Task" [ 902.004480] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.012370] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525592c4-8eb8-c516-8292-b8972746863d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.080307] env[63355]: DEBUG nova.scheduler.client.report [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.103760] env[63355]: DEBUG oslo_vmware.api [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349656, 'name': PowerOnVM_Task, 'duration_secs': 0.617142} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.104061] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.104281] env[63355]: INFO nova.compute.manager [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Took 8.97 seconds to spawn the instance on the hypervisor. [ 902.104493] env[63355]: DEBUG nova.compute.manager [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.105295] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a74912-7e2e-4fa2-a74d-caa3a9953bfb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.245849] env[63355]: DEBUG nova.compute.manager [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 902.266629] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 902.267141] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 902.267141] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 902.267320] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 902.267420] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 902.267577] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 902.267799] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 902.268373] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 902.268373] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 902.268373] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 902.268539] env[63355]: DEBUG nova.virt.hardware [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.269441] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dabef6-2ded-456c-9ac5-58db531b0550 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.278165] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024d8a68-9385-4463-aed6-e762b21a4963 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.293402] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.299034] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Creating folder: Project (4f6aae0acf84452bb9104f0647c17cdd). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 902.299352] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4183d1cf-8b42-4e79-8a09-4d8c26068743 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.309753] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Created folder: Project (4f6aae0acf84452bb9104f0647c17cdd) in parent group-v287607. [ 902.309938] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Creating folder: Instances. Parent ref: group-v287661. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 902.310187] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-879afcc4-fece-4ef7-8d3c-ddc00f7a536e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.318687] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Created folder: Instances in parent group-v287661. [ 902.318907] env[63355]: DEBUG oslo.service.loopingcall [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.319113] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 902.319314] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-77568752-dab0-422a-8e85-be7144aa48a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.334942] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.334942] env[63355]: value = "task-1349660" [ 902.334942] env[63355]: _type = "Task" [ 902.334942] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.342507] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349660, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.410898] env[63355]: DEBUG oslo_vmware.api [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349657, 'name': PowerOnVM_Task, 'duration_secs': 0.576385} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.411139] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.411352] env[63355]: INFO nova.compute.manager [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Took 6.91 seconds to spawn the instance on the hypervisor. [ 902.411539] env[63355]: DEBUG nova.compute.manager [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.412366] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63bfe73e-82af-4a9b-9770-8b68238fa43d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.514936] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525592c4-8eb8-c516-8292-b8972746863d, 'name': SearchDatastore_Task, 'duration_secs': 0.009247} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.515323] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.515622] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] b6056441-9ee4-484f-a1d2-077546f2c581/b6056441-9ee4-484f-a1d2-077546f2c581.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 902.515905] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8201ceb-aac3-476a-b411-2f2ad4fb6f67 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.523949] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for the task: (returnval){ [ 902.523949] env[63355]: value = "task-1349661" [ 902.523949] env[63355]: _type = "Task" [ 902.523949] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.532232] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349661, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.585862] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.586163] env[63355]: DEBUG nova.compute.manager [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 902.588974] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.066s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.589313] env[63355]: DEBUG nova.objects.instance [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lazy-loading 'resources' on Instance uuid 5591bb55-83d7-4301-a3f9-fde945632344 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.625781] env[63355]: INFO nova.compute.manager [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Took 32.75 seconds to build instance. [ 902.848250] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349660, 'name': CreateVM_Task, 'duration_secs': 0.397547} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.848516] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 902.849099] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.849397] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.849794] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.851054] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-925579c1-7b78-4c13-bc7e-5b10d4ec449c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.856105] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 902.856105] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ff7456-ba15-32cb-56d2-948bd9d13a37" [ 902.856105] env[63355]: _type = "Task" [ 902.856105] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.865136] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ff7456-ba15-32cb-56d2-948bd9d13a37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.931455] env[63355]: INFO nova.compute.manager [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Took 26.39 seconds to build instance. [ 903.034704] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349661, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499579} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.035052] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] b6056441-9ee4-484f-a1d2-077546f2c581/b6056441-9ee4-484f-a1d2-077546f2c581.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.035343] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.035661] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c55a2eb8-4fe3-4d13-aa14-be88f2e6864d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.042166] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for the task: (returnval){ [ 903.042166] env[63355]: value = "task-1349662" [ 903.042166] env[63355]: _type = "Task" [ 903.042166] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.051472] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349662, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.092182] env[63355]: DEBUG nova.compute.utils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 903.096426] env[63355]: DEBUG nova.compute.manager [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 903.096594] env[63355]: DEBUG nova.network.neutron [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 903.129085] env[63355]: DEBUG oslo_concurrency.lockutils [None req-53c3b060-23bc-473c-bbb7-ea5154f2208f tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.135s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.152881] env[63355]: DEBUG nova.policy [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aeb1b372d56145f782d465ef99dbd362', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea43181da5ec41d585eb2e743d036683', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 903.367132] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ff7456-ba15-32cb-56d2-948bd9d13a37, 'name': SearchDatastore_Task, 'duration_secs': 0.020553} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.369814] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.370152] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.370439] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.370633] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.370860] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.371381] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-beabb4a3-f371-4180-ad5e-c7b60d7f30fd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.381554] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.382087] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 903.382886] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5a4f308-cfbd-46b8-b0d5-976c703d48be {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.389549] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31383e64-575f-4b01-862b-086161393c85 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.399051] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 903.399051] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529b0f65-c0d1-ebfd-5205-8d83d1ede7d8" [ 903.399051] env[63355]: _type = "Task" [ 903.399051] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.402881] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c80743d-6dee-479a-90c5-9598c9c9f469 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.410143] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529b0f65-c0d1-ebfd-5205-8d83d1ede7d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009266} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.410967] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-550d4d22-53a4-4e3d-9e94-c14ab4ac9e87 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.439522] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d096a7cd-aa01-4724-b42b-f6b4e61d04f9 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.838s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.441795] env[63355]: DEBUG nova.network.neutron [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Successfully created port: 31d968dc-6934-46fe-86bd-fab6751b337c {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 903.445975] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab6444b-e433-4734-9703-e6449b67744d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.448054] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 903.448054] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c8bef4-c302-cdd5-c2db-92476bfb4807" [ 903.448054] env[63355]: _type = "Task" [ 903.448054] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.455227] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc59a18-b45d-468e-9342-83e3d2d382fe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.463098] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c8bef4-c302-cdd5-c2db-92476bfb4807, 'name': SearchDatastore_Task, 'duration_secs': 0.010284} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.463836] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.464020] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] b4b09b1d-680e-47b8-aa8a-9b3d9167824d/b4b09b1d-680e-47b8-aa8a-9b3d9167824d.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.464278] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7dffca80-461b-41c4-a18b-b5fde8f5b5fb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.475299] env[63355]: DEBUG nova.compute.provider_tree [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.480868] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 903.480868] env[63355]: value = "task-1349663" [ 903.480868] env[63355]: _type = "Task" [ 903.480868] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.490496] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349663, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.554194] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349662, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075846} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.554496] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 903.555339] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd13971-db90-44d9-9729-48446758d0f3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.578442] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] b6056441-9ee4-484f-a1d2-077546f2c581/b6056441-9ee4-484f-a1d2-077546f2c581.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.578829] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-faf4cee3-092a-427d-a099-a978daa17170 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.599339] env[63355]: DEBUG nova.compute.manager [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 903.605051] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for the task: (returnval){ [ 903.605051] env[63355]: value = "task-1349664" [ 903.605051] env[63355]: _type = "Task" [ 903.605051] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.615476] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349664, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.640396] env[63355]: DEBUG nova.compute.manager [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 903.949113] env[63355]: DEBUG nova.compute.manager [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 903.978683] env[63355]: DEBUG nova.scheduler.client.report [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.994984] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349663, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472908} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.995237] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] b4b09b1d-680e-47b8-aa8a-9b3d9167824d/b4b09b1d-680e-47b8-aa8a-9b3d9167824d.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.995506] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.995889] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90cc71e6-a6a9-4664-8939-e2abb0675a89 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.003966] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 904.003966] env[63355]: value = "task-1349665" [ 904.003966] env[63355]: _type = "Task" [ 904.003966] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.012601] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.107018] env[63355]: DEBUG nova.compute.manager [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.110466] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30593117-09ce-4f30-a69c-7138cd77eb2a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.124513] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349664, 'name': ReconfigVM_Task, 'duration_secs': 0.402075} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.126372] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Reconfigured VM instance instance-00000041 to attach disk [datastore2] b6056441-9ee4-484f-a1d2-077546f2c581/b6056441-9ee4-484f-a1d2-077546f2c581.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.129560] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cacd83b1-b546-42cd-a248-9cc7dd2c1c2a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.133196] env[63355]: INFO nova.compute.manager [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Rescuing [ 904.133433] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "refresh_cache-8e3ec9d3-bc22-4e39-ad7c-93268dd59020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.133615] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "refresh_cache-8e3ec9d3-bc22-4e39-ad7c-93268dd59020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.133794] env[63355]: DEBUG nova.network.neutron [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.137026] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for the task: (returnval){ [ 904.137026] env[63355]: value = "task-1349666" [ 904.137026] env[63355]: _type = "Task" [ 904.137026] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.148231] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349666, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.165605] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.470398] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.488860] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.492475] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.797s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.494618] env[63355]: INFO nova.compute.claims [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.512312] env[63355]: INFO nova.scheduler.client.report [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted allocations for instance 5591bb55-83d7-4301-a3f9-fde945632344 [ 904.518285] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079221} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.522078] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.523522] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509bd082-28d1-407a-b78d-b48f3303325e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.556726] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] b4b09b1d-680e-47b8-aa8a-9b3d9167824d/b4b09b1d-680e-47b8-aa8a-9b3d9167824d.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.558185] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75ccdea1-3063-4b2b-8d01-0c29c420a31e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.591129] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 904.591129] env[63355]: value = "task-1349667" [ 904.591129] env[63355]: _type = "Task" [ 904.591129] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.604286] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349667, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.615606] env[63355]: DEBUG nova.compute.manager [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 904.634240] env[63355]: INFO nova.compute.manager [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] instance snapshotting [ 904.639099] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-307dc393-3066-4c16-9ebd-b9a1913d8d17 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.644243] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.644701] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.644701] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.644890] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.645106] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.645275] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.645573] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.645793] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.645915] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.646100] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.646282] env[63355]: DEBUG nova.virt.hardware [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.650383] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13f18e6-5c5b-4a82-931c-7025f913a2f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.661832] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349666, 'name': Rename_Task, 'duration_secs': 0.262821} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.675181] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.676600] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4de073-c6a5-4955-84a1-32fab690d1dc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.683421] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2b4c5e4-d413-412d-8804-eb62f6456cea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.685484] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9c2de7-317e-4e8d-9d3d-35c06d2c0738 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.705237] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for the task: (returnval){ [ 904.705237] env[63355]: value = "task-1349668" [ 904.705237] env[63355]: _type = "Task" [ 904.705237] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.717701] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349668, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.026326] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6a23847c-90df-4bef-8fff-ef99709f7a36 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "5591bb55-83d7-4301-a3f9-fde945632344" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.493s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.102426] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349667, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.146884] env[63355]: DEBUG nova.network.neutron [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Updating instance_info_cache with network_info: [{"id": "b134b9dc-a12b-48b4-8051-39ec142ce649", "address": "fa:16:3e:30:2c:29", "network": {"id": "26b0dabd-e662-4b47-bcd4-a50e05f0a52d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-825312565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1efb1730acbf4ba5a19bd150bf99d48f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb134b9dc-a1", "ovs_interfaceid": "b134b9dc-a12b-48b4-8051-39ec142ce649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.214223] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 905.214508] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ca7f7fc9-5e7f-43f1-8ac9-19ef6c05b72e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.225062] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349668, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.227046] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 905.227046] env[63355]: value = "task-1349669" [ 905.227046] env[63355]: _type = "Task" [ 905.227046] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.238313] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349669, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.310103] env[63355]: DEBUG nova.network.neutron [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Successfully updated port: 31d968dc-6934-46fe-86bd-fab6751b337c {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 905.319465] env[63355]: DEBUG nova.compute.manager [req-e1ca3b9f-453d-4f47-8fec-d74ac41b79c1 req-37abcb86-4f12-44e6-a261-7adc558d0218 service nova] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Received event network-vif-plugged-31d968dc-6934-46fe-86bd-fab6751b337c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.319465] env[63355]: DEBUG oslo_concurrency.lockutils [req-e1ca3b9f-453d-4f47-8fec-d74ac41b79c1 req-37abcb86-4f12-44e6-a261-7adc558d0218 service nova] Acquiring lock "2539a79e-01c0-4e0c-aa66-8784441c6fda-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.319465] env[63355]: DEBUG oslo_concurrency.lockutils [req-e1ca3b9f-453d-4f47-8fec-d74ac41b79c1 req-37abcb86-4f12-44e6-a261-7adc558d0218 service nova] Lock "2539a79e-01c0-4e0c-aa66-8784441c6fda-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.319677] env[63355]: DEBUG oslo_concurrency.lockutils [req-e1ca3b9f-453d-4f47-8fec-d74ac41b79c1 req-37abcb86-4f12-44e6-a261-7adc558d0218 service nova] Lock "2539a79e-01c0-4e0c-aa66-8784441c6fda-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.320048] env[63355]: DEBUG nova.compute.manager [req-e1ca3b9f-453d-4f47-8fec-d74ac41b79c1 req-37abcb86-4f12-44e6-a261-7adc558d0218 service nova] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] No waiting events found dispatching network-vif-plugged-31d968dc-6934-46fe-86bd-fab6751b337c {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 905.320048] env[63355]: WARNING nova.compute.manager [req-e1ca3b9f-453d-4f47-8fec-d74ac41b79c1 req-37abcb86-4f12-44e6-a261-7adc558d0218 service nova] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Received unexpected event network-vif-plugged-31d968dc-6934-46fe-86bd-fab6751b337c for instance with vm_state building and task_state spawning. [ 905.602622] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349667, 'name': ReconfigVM_Task, 'duration_secs': 0.553642} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.604965] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Reconfigured VM instance instance-00000042 to attach disk [datastore1] b4b09b1d-680e-47b8-aa8a-9b3d9167824d/b4b09b1d-680e-47b8-aa8a-9b3d9167824d.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.605823] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b414239-7889-438d-b077-5ffb94008c8a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.613295] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 905.613295] env[63355]: value = "task-1349670" [ 905.613295] env[63355]: _type = "Task" [ 905.613295] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.625253] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349670, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.649581] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "refresh_cache-8e3ec9d3-bc22-4e39-ad7c-93268dd59020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.720282] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349668, 'name': PowerOnVM_Task} progress is 81%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.740425] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349669, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.784075] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d71079-7600-4fbf-94e6-22890739935a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.792569] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54103233-e6da-4cf4-8a7b-52fc3af0c78d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.826192] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquiring lock "refresh_cache-2539a79e-01c0-4e0c-aa66-8784441c6fda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.826359] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquired lock "refresh_cache-2539a79e-01c0-4e0c-aa66-8784441c6fda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.826535] env[63355]: DEBUG nova.network.neutron [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 905.828975] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb92427-0704-42d1-af82-9fa5f443486e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.838498] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb766f9-3662-4025-858d-54afd17272e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.857720] env[63355]: DEBUG nova.compute.provider_tree [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.124109] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349670, 'name': Rename_Task, 'duration_secs': 0.270249} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.124463] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.124751] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fec4c659-1454-4d95-b803-7eb9a2825d0e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.132840] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 906.132840] env[63355]: value = "task-1349671" [ 906.132840] env[63355]: _type = "Task" [ 906.132840] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.143013] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.176810] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.177156] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-feb19583-cf65-416c-af9f-ad4614717040 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.186460] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 906.186460] env[63355]: value = "task-1349672" [ 906.186460] env[63355]: _type = "Task" [ 906.186460] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.195436] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349672, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.219655] env[63355]: DEBUG oslo_vmware.api [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349668, 'name': PowerOnVM_Task, 'duration_secs': 1.272701} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.220414] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 906.220414] env[63355]: INFO nova.compute.manager [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Took 8.36 seconds to spawn the instance on the hypervisor. [ 906.220574] env[63355]: DEBUG nova.compute.manager [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.221320] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b6274c-581f-45e8-9f10-e16dade0358f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.240858] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349669, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.361233] env[63355]: DEBUG nova.scheduler.client.report [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 906.365811] env[63355]: DEBUG nova.network.neutron [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 906.469990] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "fc6a99e6-5319-47d3-8175-770c06d3e325" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.470250] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "fc6a99e6-5319-47d3-8175-770c06d3e325" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.491121] env[63355]: DEBUG oslo_vmware.rw_handles [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529066c3-c3ca-f757-fb1d-9dc2d1b34bb0/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 906.492126] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbae01f7-9ed7-411e-845f-ac53e1e70587 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.498752] env[63355]: DEBUG oslo_vmware.rw_handles [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529066c3-c3ca-f757-fb1d-9dc2d1b34bb0/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 906.498941] env[63355]: ERROR oslo_vmware.rw_handles [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529066c3-c3ca-f757-fb1d-9dc2d1b34bb0/disk-0.vmdk due to incomplete transfer. [ 906.499199] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5ece4764-2bd5-4c20-81b7-006930396772 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.507040] env[63355]: DEBUG oslo_vmware.rw_handles [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529066c3-c3ca-f757-fb1d-9dc2d1b34bb0/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 906.507249] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Uploaded image 2584bf77-1454-4062-a175-2abf8f10d255 to the Glance image server {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 906.509580] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 906.509839] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bc0b5112-5853-4540-aec8-69179485177f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.517601] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 906.517601] env[63355]: value = "task-1349673" [ 906.517601] env[63355]: _type = "Task" [ 906.517601] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.518863] env[63355]: DEBUG nova.network.neutron [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Updating instance_info_cache with network_info: [{"id": "31d968dc-6934-46fe-86bd-fab6751b337c", "address": "fa:16:3e:2f:9a:4d", "network": {"id": "fdad4482-3471-4d5b-94fb-10718bba8f68", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1155876767-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea43181da5ec41d585eb2e743d036683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31d968dc-69", "ovs_interfaceid": "31d968dc-6934-46fe-86bd-fab6751b337c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.529543] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349673, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.643818] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349671, 'name': PowerOnVM_Task} progress is 80%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.697081] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349672, 'name': PowerOffVM_Task, 'duration_secs': 0.20172} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.697371] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.698195] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab36033-1433-41f6-a4c6-fab97b95d9b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.717089] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce7af52-5c37-44e3-90e3-757183fac77d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.742050] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349669, 'name': CreateSnapshot_Task, 'duration_secs': 1.122484} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.744152] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 906.745142] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba398d3-29a6-4606-9cdb-90b49ff65f63 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.752315] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.752608] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a8aa719-71af-4411-8a0d-2ebe0bc96f72 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.757443] env[63355]: INFO nova.compute.manager [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Took 26.94 seconds to build instance. [ 906.766767] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 906.766767] env[63355]: value = "task-1349674" [ 906.766767] env[63355]: _type = "Task" [ 906.766767] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.776142] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349674, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.869351] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.377s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.869895] env[63355]: DEBUG nova.compute.manager [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 906.872468] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.615s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.873906] env[63355]: INFO nova.compute.claims [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.024429] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Releasing lock "refresh_cache-2539a79e-01c0-4e0c-aa66-8784441c6fda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.024787] env[63355]: DEBUG nova.compute.manager [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Instance network_info: |[{"id": "31d968dc-6934-46fe-86bd-fab6751b337c", "address": "fa:16:3e:2f:9a:4d", "network": {"id": "fdad4482-3471-4d5b-94fb-10718bba8f68", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1155876767-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea43181da5ec41d585eb2e743d036683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31d968dc-69", "ovs_interfaceid": "31d968dc-6934-46fe-86bd-fab6751b337c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 907.025372] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:9a:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1470a3f8-be8a-4339-8a6f-9519366f32e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31d968dc-6934-46fe-86bd-fab6751b337c', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.033030] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Creating folder: Project (ea43181da5ec41d585eb2e743d036683). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 907.033861] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce1d6fe6-4ec5-4498-adcb-7fed978bf2b7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.039279] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349673, 'name': Destroy_Task} progress is 33%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.049877] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Created folder: Project (ea43181da5ec41d585eb2e743d036683) in parent group-v287607. [ 907.050105] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Creating folder: Instances. Parent ref: group-v287665. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 907.050361] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4c9f091-e6de-4326-907b-ac767c9b7bf4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.053723] env[63355]: DEBUG oslo_concurrency.lockutils [None req-76ee6292-3039-44f2-b579-c71b52a0c2a4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquiring lock "interface-b6056441-9ee4-484f-a1d2-077546f2c581-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.053962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-76ee6292-3039-44f2-b579-c71b52a0c2a4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "interface-b6056441-9ee4-484f-a1d2-077546f2c581-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.054304] env[63355]: DEBUG nova.objects.instance [None req-76ee6292-3039-44f2-b579-c71b52a0c2a4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lazy-loading 'flavor' on Instance uuid b6056441-9ee4-484f-a1d2-077546f2c581 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.062530] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Created folder: Instances in parent group-v287665. [ 907.062754] env[63355]: DEBUG oslo.service.loopingcall [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.062966] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 907.063283] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7538a817-d505-41f9-8624-a7c6937b5788 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.085132] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 907.085132] env[63355]: value = "task-1349677" [ 907.085132] env[63355]: _type = "Task" [ 907.085132] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.096069] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349677, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.143832] env[63355]: DEBUG oslo_vmware.api [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349671, 'name': PowerOnVM_Task, 'duration_secs': 0.973498} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.144278] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.144443] env[63355]: INFO nova.compute.manager [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Took 4.90 seconds to spawn the instance on the hypervisor. [ 907.144629] env[63355]: DEBUG nova.compute.manager [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.145463] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01ab6b5-3f51-4371-9065-863e5dd7397f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.261180] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2c59ae7d-d828-41fc-baf0-5bed639f5224 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "b6056441-9ee4-484f-a1d2-077546f2c581" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.624s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.269537] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 907.269938] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d2522ac4-7bfc-4189-804d-3da2b9113129 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.284532] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] VM already powered off {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 907.284764] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.285013] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.285200] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.285385] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.286012] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 907.286012] env[63355]: value = "task-1349678" [ 907.286012] env[63355]: _type = "Task" [ 907.286012] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.286242] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d625be07-eff2-40c7-beb8-2cae3ff2b701 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.297376] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349678, 'name': CloneVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.298683] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.298876] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 907.299935] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40f90857-75f0-4a1e-8e2e-d4644dd4cd3e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.307507] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 907.307507] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524d9abe-e09f-636f-f3df-307dff899661" [ 907.307507] env[63355]: _type = "Task" [ 907.307507] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.316829] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524d9abe-e09f-636f-f3df-307dff899661, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.379079] env[63355]: DEBUG nova.compute.utils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.382312] env[63355]: DEBUG nova.compute.manager [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 907.382532] env[63355]: DEBUG nova.network.neutron [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 907.437426] env[63355]: DEBUG nova.policy [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '81c301c95dfa4b409dc3e11917209a49', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '214e0560beec42fea860b59ae67d374a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 907.533631] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349673, 'name': Destroy_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.544707] env[63355]: DEBUG nova.compute.manager [req-0cd518fd-883e-4869-95f0-f322bc86e0fe req-7676bfaa-bcf6-4d2c-a7b6-f00e2825df0b service nova] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Received event network-changed-31d968dc-6934-46fe-86bd-fab6751b337c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.544911] env[63355]: DEBUG nova.compute.manager [req-0cd518fd-883e-4869-95f0-f322bc86e0fe req-7676bfaa-bcf6-4d2c-a7b6-f00e2825df0b service nova] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Refreshing instance network info cache due to event network-changed-31d968dc-6934-46fe-86bd-fab6751b337c. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 907.545136] env[63355]: DEBUG oslo_concurrency.lockutils [req-0cd518fd-883e-4869-95f0-f322bc86e0fe req-7676bfaa-bcf6-4d2c-a7b6-f00e2825df0b service nova] Acquiring lock "refresh_cache-2539a79e-01c0-4e0c-aa66-8784441c6fda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.545322] env[63355]: DEBUG oslo_concurrency.lockutils [req-0cd518fd-883e-4869-95f0-f322bc86e0fe req-7676bfaa-bcf6-4d2c-a7b6-f00e2825df0b service nova] Acquired lock "refresh_cache-2539a79e-01c0-4e0c-aa66-8784441c6fda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.545577] env[63355]: DEBUG nova.network.neutron [req-0cd518fd-883e-4869-95f0-f322bc86e0fe req-7676bfaa-bcf6-4d2c-a7b6-f00e2825df0b service nova] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Refreshing network info cache for port 31d968dc-6934-46fe-86bd-fab6751b337c {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 907.558230] env[63355]: DEBUG nova.objects.instance [None req-76ee6292-3039-44f2-b579-c71b52a0c2a4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lazy-loading 'pci_requests' on Instance uuid b6056441-9ee4-484f-a1d2-077546f2c581 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.597634] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349677, 'name': CreateVM_Task, 'duration_secs': 0.436593} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.597851] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 907.598514] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.598687] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.599069] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 907.599356] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9abe19dd-492b-4936-8b5d-2b6bb243a5b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.607151] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for the task: (returnval){ [ 907.607151] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528d672b-0641-9b15-0f9d-76be6ee9ca60" [ 907.607151] env[63355]: _type = "Task" [ 907.607151] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.619454] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528d672b-0641-9b15-0f9d-76be6ee9ca60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.664154] env[63355]: INFO nova.compute.manager [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Took 24.11 seconds to build instance. [ 907.700830] env[63355]: DEBUG nova.network.neutron [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Successfully created port: 00f30816-4a44-44bc-9eb3-ca224f3da164 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 907.771381] env[63355]: DEBUG nova.compute.manager [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 907.800206] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349678, 'name': CloneVM_Task} progress is 93%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.819892] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524d9abe-e09f-636f-f3df-307dff899661, 'name': SearchDatastore_Task, 'duration_secs': 0.011221} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.820807] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33d5b16c-4db8-4bb8-abbb-e403e392b984 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.829020] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 907.829020] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52dc0598-f9b9-c444-e889-6b768ed83baf" [ 907.829020] env[63355]: _type = "Task" [ 907.829020] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.836400] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52dc0598-f9b9-c444-e889-6b768ed83baf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.884219] env[63355]: DEBUG nova.compute.manager [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 908.035545] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349673, 'name': Destroy_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.060450] env[63355]: DEBUG nova.objects.base [None req-76ee6292-3039-44f2-b579-c71b52a0c2a4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 908.060680] env[63355]: DEBUG nova.network.neutron [None req-76ee6292-3039-44f2-b579-c71b52a0c2a4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 908.120428] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528d672b-0641-9b15-0f9d-76be6ee9ca60, 'name': SearchDatastore_Task, 'duration_secs': 0.010302} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.122524] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.122790] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 908.123096] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.123194] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.123378] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 908.125944] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a96a984-17b0-41ae-9415-a6b3bff8c6fa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.139222] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.139222] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 908.139222] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e05ad11d-f6dc-4a3f-8027-d4bae5860314 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.150806] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for the task: (returnval){ [ 908.150806] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d0a7f6-9ed0-7568-941e-9eb3e62f1ee9" [ 908.150806] env[63355]: _type = "Task" [ 908.150806] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.159762] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d0a7f6-9ed0-7568-941e-9eb3e62f1ee9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.168640] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ddd19b7f-055a-4326-9fcb-ec8adc9fabf5 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "b4b09b1d-680e-47b8-aa8a-9b3d9167824d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.113s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.187050] env[63355]: DEBUG oslo_concurrency.lockutils [None req-76ee6292-3039-44f2-b579-c71b52a0c2a4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "interface-b6056441-9ee4-484f-a1d2-077546f2c581-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.133s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.194954] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a453ed13-1517-4dbb-a448-a47be797593f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.206442] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c7bc1c-509c-40b6-a5f8-d9294ad356f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.248063] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20400e04-e15d-451a-ab8c-b69d458c4c1c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.256565] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d39d6c3-de68-4be2-8e85-ddac2e9c7417 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.270978] env[63355]: DEBUG nova.compute.provider_tree [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.294075] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.302810] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349678, 'name': CloneVM_Task} progress is 93%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.320676] env[63355]: DEBUG nova.network.neutron [req-0cd518fd-883e-4869-95f0-f322bc86e0fe req-7676bfaa-bcf6-4d2c-a7b6-f00e2825df0b service nova] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Updated VIF entry in instance network info cache for port 31d968dc-6934-46fe-86bd-fab6751b337c. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 908.321052] env[63355]: DEBUG nova.network.neutron [req-0cd518fd-883e-4869-95f0-f322bc86e0fe req-7676bfaa-bcf6-4d2c-a7b6-f00e2825df0b service nova] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Updating instance_info_cache with network_info: [{"id": "31d968dc-6934-46fe-86bd-fab6751b337c", "address": "fa:16:3e:2f:9a:4d", "network": {"id": "fdad4482-3471-4d5b-94fb-10718bba8f68", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1155876767-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea43181da5ec41d585eb2e743d036683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31d968dc-69", "ovs_interfaceid": "31d968dc-6934-46fe-86bd-fab6751b337c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.338026] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52dc0598-f9b9-c444-e889-6b768ed83baf, 'name': SearchDatastore_Task, 'duration_secs': 0.033882} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.338302] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.338562] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 8e3ec9d3-bc22-4e39-ad7c-93268dd59020/84ca0bb3-9916-4cea-9399-75af77b0558e-rescue.vmdk. {{(pid=63355) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 908.338847] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a563663-4db9-4a51-bd02-e0b74da7db15 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.348379] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 908.348379] env[63355]: value = "task-1349679" [ 908.348379] env[63355]: _type = "Task" [ 908.348379] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.357207] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349679, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.380558] env[63355]: INFO nova.compute.manager [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Rebuilding instance [ 908.424839] env[63355]: DEBUG nova.compute.manager [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.425752] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7860b0f9-4830-412d-aed4-700a6c7dd141 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.535299] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349673, 'name': Destroy_Task, 'duration_secs': 1.877301} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.535926] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Destroyed the VM [ 908.536375] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 908.536741] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-35cb96cd-62be-471a-8f81-9ade103505b8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.549041] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 908.549041] env[63355]: value = "task-1349680" [ 908.549041] env[63355]: _type = "Task" [ 908.549041] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.560634] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349680, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.664673] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d0a7f6-9ed0-7568-941e-9eb3e62f1ee9, 'name': SearchDatastore_Task, 'duration_secs': 0.013815} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.665773] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3eeebaef-2078-4238-9bc4-a1e22196652a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.670291] env[63355]: DEBUG nova.compute.manager [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.676059] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for the task: (returnval){ [ 908.676059] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52089f2b-6d27-7500-847c-76c21b9219b2" [ 908.676059] env[63355]: _type = "Task" [ 908.676059] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.687575] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52089f2b-6d27-7500-847c-76c21b9219b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.773788] env[63355]: DEBUG nova.scheduler.client.report [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.802915] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349678, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.823792] env[63355]: DEBUG oslo_concurrency.lockutils [req-0cd518fd-883e-4869-95f0-f322bc86e0fe req-7676bfaa-bcf6-4d2c-a7b6-f00e2825df0b service nova] Releasing lock "refresh_cache-2539a79e-01c0-4e0c-aa66-8784441c6fda" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.859855] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349679, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510275} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.859855] env[63355]: INFO nova.virt.vmwareapi.ds_util [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 8e3ec9d3-bc22-4e39-ad7c-93268dd59020/84ca0bb3-9916-4cea-9399-75af77b0558e-rescue.vmdk. [ 908.859855] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999bdaf3-6362-469e-942f-77a49dc87fac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.886603] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 8e3ec9d3-bc22-4e39-ad7c-93268dd59020/84ca0bb3-9916-4cea-9399-75af77b0558e-rescue.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.886883] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-455f7ddf-6659-4a73-a8e3-90ef4ecfabd6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.901602] env[63355]: DEBUG nova.compute.manager [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 908.911318] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 908.911318] env[63355]: value = "task-1349681" [ 908.911318] env[63355]: _type = "Task" [ 908.911318] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.921780] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349681, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.924725] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 908.924970] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 908.925149] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.925419] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 908.925541] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.925704] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 908.925932] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 908.926111] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 908.926283] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 908.926447] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 908.926703] env[63355]: DEBUG nova.virt.hardware [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.927571] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e99f31-4d78-4336-baf0-26548fd90d80 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.936871] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44482978-a0a1-4e31-a317-4988110469ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.941918] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.942275] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40b3f736-0c36-475a-b496-1de7d0f5c9f2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.957619] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 908.957619] env[63355]: value = "task-1349682" [ 908.957619] env[63355]: _type = "Task" [ 908.957619] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.968408] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349682, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.062266] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349680, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.190311] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52089f2b-6d27-7500-847c-76c21b9219b2, 'name': SearchDatastore_Task, 'duration_secs': 0.020182} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.190664] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.190989] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 2539a79e-01c0-4e0c-aa66-8784441c6fda/2539a79e-01c0-4e0c-aa66-8784441c6fda.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 909.191428] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0a943539-ed88-4be0-88d1-e0425d6a9375 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.202685] env[63355]: DEBUG nova.compute.manager [req-63c0d8b8-79c7-4671-b209-e45bcc81d11b req-64febb76-5f1c-4569-923d-cb499a79225b service nova] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Received event network-vif-plugged-00f30816-4a44-44bc-9eb3-ca224f3da164 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.202898] env[63355]: DEBUG oslo_concurrency.lockutils [req-63c0d8b8-79c7-4671-b209-e45bcc81d11b req-64febb76-5f1c-4569-923d-cb499a79225b service nova] Acquiring lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.203073] env[63355]: DEBUG oslo_concurrency.lockutils [req-63c0d8b8-79c7-4671-b209-e45bcc81d11b req-64febb76-5f1c-4569-923d-cb499a79225b service nova] Lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.203252] env[63355]: DEBUG oslo_concurrency.lockutils [req-63c0d8b8-79c7-4671-b209-e45bcc81d11b req-64febb76-5f1c-4569-923d-cb499a79225b service nova] Lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.203471] env[63355]: DEBUG nova.compute.manager [req-63c0d8b8-79c7-4671-b209-e45bcc81d11b req-64febb76-5f1c-4569-923d-cb499a79225b service nova] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] No waiting events found dispatching network-vif-plugged-00f30816-4a44-44bc-9eb3-ca224f3da164 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 909.203663] env[63355]: WARNING nova.compute.manager [req-63c0d8b8-79c7-4671-b209-e45bcc81d11b req-64febb76-5f1c-4569-923d-cb499a79225b service nova] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Received unexpected event network-vif-plugged-00f30816-4a44-44bc-9eb3-ca224f3da164 for instance with vm_state building and task_state spawning. [ 909.204903] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.208130] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for the task: (returnval){ [ 909.208130] env[63355]: value = "task-1349683" [ 909.208130] env[63355]: _type = "Task" [ 909.208130] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.218688] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349683, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.279728] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.280479] env[63355]: DEBUG nova.compute.manager [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 909.283594] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.698s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.285951] env[63355]: INFO nova.compute.claims [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.301833] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349678, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.424935] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349681, 'name': ReconfigVM_Task, 'duration_secs': 0.448668} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.425325] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 8e3ec9d3-bc22-4e39-ad7c-93268dd59020/84ca0bb3-9916-4cea-9399-75af77b0558e-rescue.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.426469] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa695a05-0fb9-456a-a38e-32229f4fea60 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.455252] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-455e5dd2-3c42-40cb-8a0d-37a6103d8eb5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.476842] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349682, 'name': PowerOffVM_Task, 'duration_secs': 0.279717} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.478634] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.478906] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 909.479354] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 909.479354] env[63355]: value = "task-1349684" [ 909.479354] env[63355]: _type = "Task" [ 909.479354] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.480114] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e26549-ae27-4290-b4da-17bd54f1c70f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.497473] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349684, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.497868] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 909.498158] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f84f733b-551b-4ba6-b5cd-2119a44c3609 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.532325] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.532554] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.532751] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Deleting the datastore file [datastore1] b4b09b1d-680e-47b8-aa8a-9b3d9167824d {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.533051] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7e8b1c0-6dd4-4984-8c5a-d2f10730b0a8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.540800] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 909.540800] env[63355]: value = "task-1349686" [ 909.540800] env[63355]: _type = "Task" [ 909.540800] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.551183] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349686, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.560321] env[63355]: DEBUG oslo_vmware.api [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349680, 'name': RemoveSnapshot_Task, 'duration_secs': 0.946268} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.560678] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 909.560978] env[63355]: INFO nova.compute.manager [None req-d9a0e24f-840a-4098-9b7c-31aea289ed3b tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Took 15.11 seconds to snapshot the instance on the hypervisor. [ 909.683018] env[63355]: DEBUG nova.network.neutron [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Successfully updated port: 00f30816-4a44-44bc-9eb3-ca224f3da164 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.720232] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349683, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507648} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.720610] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 2539a79e-01c0-4e0c-aa66-8784441c6fda/2539a79e-01c0-4e0c-aa66-8784441c6fda.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 909.720780] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.721032] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f6a64c35-37bb-4836-b46d-aaed0a62479e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.729319] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for the task: (returnval){ [ 909.729319] env[63355]: value = "task-1349687" [ 909.729319] env[63355]: _type = "Task" [ 909.729319] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.738930] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349687, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.782593] env[63355]: DEBUG nova.compute.manager [req-c3e40b7c-ec41-45f0-b065-321acdce7b27 req-01a694e8-fa06-4cf0-b547-d9c98b29f15a service nova] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Received event network-changed-00f30816-4a44-44bc-9eb3-ca224f3da164 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.783096] env[63355]: DEBUG nova.compute.manager [req-c3e40b7c-ec41-45f0-b065-321acdce7b27 req-01a694e8-fa06-4cf0-b547-d9c98b29f15a service nova] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Refreshing instance network info cache due to event network-changed-00f30816-4a44-44bc-9eb3-ca224f3da164. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.783174] env[63355]: DEBUG oslo_concurrency.lockutils [req-c3e40b7c-ec41-45f0-b065-321acdce7b27 req-01a694e8-fa06-4cf0-b547-d9c98b29f15a service nova] Acquiring lock "refresh_cache-7c139710-d8d9-4cd7-bec0-6e021d3b2e68" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.783322] env[63355]: DEBUG oslo_concurrency.lockutils [req-c3e40b7c-ec41-45f0-b065-321acdce7b27 req-01a694e8-fa06-4cf0-b547-d9c98b29f15a service nova] Acquired lock "refresh_cache-7c139710-d8d9-4cd7-bec0-6e021d3b2e68" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.783874] env[63355]: DEBUG nova.network.neutron [req-c3e40b7c-ec41-45f0-b065-321acdce7b27 req-01a694e8-fa06-4cf0-b547-d9c98b29f15a service nova] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Refreshing network info cache for port 00f30816-4a44-44bc-9eb3-ca224f3da164 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 909.789139] env[63355]: DEBUG nova.compute.utils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.790755] env[63355]: DEBUG nova.compute.manager [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.790942] env[63355]: DEBUG nova.network.neutron [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 909.806855] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349678, 'name': CloneVM_Task, 'duration_secs': 2.347749} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.807200] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Created linked-clone VM from snapshot [ 909.808018] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7af22a-86d9-4eed-a42b-c472b24fd85e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.817701] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Uploading image f755e516-c1d3-4618-aedf-9caa138936ff {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 909.843611] env[63355]: DEBUG nova.policy [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3fd9a173cf6d4f2c8da17df12a9f3779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87a296f0ecd94879ab2336899fc575bf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.847228] env[63355]: DEBUG oslo_vmware.rw_handles [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 909.847228] env[63355]: value = "vm-287668" [ 909.847228] env[63355]: _type = "VirtualMachine" [ 909.847228] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 909.847796] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-06a75886-04f3-42a4-b06a-c7b85e521b0d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.858697] env[63355]: DEBUG oslo_vmware.rw_handles [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lease: (returnval){ [ 909.858697] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52033748-270c-8e59-22c1-8c836fd9ccee" [ 909.858697] env[63355]: _type = "HttpNfcLease" [ 909.858697] env[63355]: } obtained for exporting VM: (result){ [ 909.858697] env[63355]: value = "vm-287668" [ 909.858697] env[63355]: _type = "VirtualMachine" [ 909.858697] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 909.858974] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the lease: (returnval){ [ 909.858974] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52033748-270c-8e59-22c1-8c836fd9ccee" [ 909.858974] env[63355]: _type = "HttpNfcLease" [ 909.858974] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 909.866589] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 909.866589] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52033748-270c-8e59-22c1-8c836fd9ccee" [ 909.866589] env[63355]: _type = "HttpNfcLease" [ 909.866589] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 909.994153] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349684, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.053895] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349686, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247983} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.053895] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.053895] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.053895] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.101841] env[63355]: DEBUG nova.network.neutron [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Successfully created port: 62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 910.148007] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquiring lock "b6056441-9ee4-484f-a1d2-077546f2c581" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.148316] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "b6056441-9ee4-484f-a1d2-077546f2c581" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.148540] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquiring lock "b6056441-9ee4-484f-a1d2-077546f2c581-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.149196] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "b6056441-9ee4-484f-a1d2-077546f2c581-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.149196] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "b6056441-9ee4-484f-a1d2-077546f2c581-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.150959] env[63355]: INFO nova.compute.manager [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Terminating instance [ 910.152756] env[63355]: DEBUG nova.compute.manager [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 910.152951] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 910.153776] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fff2eb1-2364-45b7-b662-98ac5ea6d28f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.162479] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.162707] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-545f88fc-af9a-4453-9e6d-1794ef6afded {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.169375] env[63355]: DEBUG oslo_vmware.api [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for the task: (returnval){ [ 910.169375] env[63355]: value = "task-1349689" [ 910.169375] env[63355]: _type = "Task" [ 910.169375] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.177983] env[63355]: DEBUG oslo_vmware.api [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349689, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.185640] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquiring lock "refresh_cache-7c139710-d8d9-4cd7-bec0-6e021d3b2e68" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.244149] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349687, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.297549] env[63355]: DEBUG nova.compute.manager [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 910.324705] env[63355]: DEBUG nova.network.neutron [req-c3e40b7c-ec41-45f0-b065-321acdce7b27 req-01a694e8-fa06-4cf0-b547-d9c98b29f15a service nova] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 910.369876] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 910.369876] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52033748-270c-8e59-22c1-8c836fd9ccee" [ 910.369876] env[63355]: _type = "HttpNfcLease" [ 910.369876] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 910.370191] env[63355]: DEBUG oslo_vmware.rw_handles [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 910.370191] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52033748-270c-8e59-22c1-8c836fd9ccee" [ 910.370191] env[63355]: _type = "HttpNfcLease" [ 910.370191] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 910.370949] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84a5e51-70ad-4e7d-83e4-e8519a9d058a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.382672] env[63355]: DEBUG oslo_vmware.rw_handles [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521e74fe-5003-7980-15f1-e31689e12e7b/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 910.382863] env[63355]: DEBUG oslo_vmware.rw_handles [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521e74fe-5003-7980-15f1-e31689e12e7b/disk-0.vmdk for reading. {{(pid=63355) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 910.483183] env[63355]: DEBUG nova.network.neutron [req-c3e40b7c-ec41-45f0-b065-321acdce7b27 req-01a694e8-fa06-4cf0-b547-d9c98b29f15a service nova] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.494223] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349684, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.650272] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dabdad81-eff1-49eb-95b2-10ece1095a1a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.658070] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e34ddda-c909-4620-908b-0fd53e753519 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.669012] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1acdbf7a-b977-4787-af65-8da2a9dc5880 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.707036] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7d801c-d5b7-40c6-bbbd-e7908d3e8709 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.714762] env[63355]: DEBUG oslo_vmware.api [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349689, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.723926] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2283a32a-fb7a-460e-b5c0-261adeaf4c19 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.753016] env[63355]: DEBUG nova.compute.provider_tree [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 910.766023] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349687, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.024101} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.766023] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 910.766023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67b1c66-de4e-489a-812f-2be72123b39c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.788618] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 2539a79e-01c0-4e0c-aa66-8784441c6fda/2539a79e-01c0-4e0c-aa66-8784441c6fda.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 910.789711] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad565d74-b485-4223-938a-3eb01657daa2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.814708] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for the task: (returnval){ [ 910.814708] env[63355]: value = "task-1349690" [ 910.814708] env[63355]: _type = "Task" [ 910.814708] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.825018] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349690, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.990739] env[63355]: DEBUG oslo_concurrency.lockutils [req-c3e40b7c-ec41-45f0-b065-321acdce7b27 req-01a694e8-fa06-4cf0-b547-d9c98b29f15a service nova] Releasing lock "refresh_cache-7c139710-d8d9-4cd7-bec0-6e021d3b2e68" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.991252] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquired lock "refresh_cache-7c139710-d8d9-4cd7-bec0-6e021d3b2e68" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.991476] env[63355]: DEBUG nova.network.neutron [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.001521] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349684, 'name': ReconfigVM_Task, 'duration_secs': 1.250517} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.002144] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 911.002469] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ccebddc-25ef-448c-8947-a81856313224 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.011947] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 911.011947] env[63355]: value = "task-1349691" [ 911.011947] env[63355]: _type = "Task" [ 911.011947] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.021814] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349691, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.093149] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.093946] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.094204] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.094604] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.094850] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.095036] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.095412] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.095971] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.096622] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.096955] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.097225] env[63355]: DEBUG nova.virt.hardware [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.098204] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c30f032-555d-4f94-b1b6-7df88384a0eb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.109712] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e50d817-018f-4e83-b2f3-4b161cf6e5ce {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.126995] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Instance VIF info [] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.134112] env[63355]: DEBUG oslo.service.loopingcall [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.134404] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 911.134634] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3df1f6e-0e06-4fa3-8d5d-fea79edf5b91 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.156635] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.156635] env[63355]: value = "task-1349692" [ 911.156635] env[63355]: _type = "Task" [ 911.156635] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.167815] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349692, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.189668] env[63355]: DEBUG oslo_vmware.api [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349689, 'name': PowerOffVM_Task, 'duration_secs': 0.763944} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.189999] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 911.190766] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 911.190766] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8feecd7c-4c95-41ae-8666-63beb72b3ee6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.273122] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 911.273657] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 911.274181] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Deleting the datastore file [datastore2] b6056441-9ee4-484f-a1d2-077546f2c581 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.274490] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edee2908-b978-4746-aacb-c9a3c18ddf7f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.277567] env[63355]: ERROR nova.scheduler.client.report [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [req-7629111b-3fd7-4a12-8bd5-4b3d75c8c73e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 47abb610-db7e-4770-911d-187dd075ef8b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7629111b-3fd7-4a12-8bd5-4b3d75c8c73e"}]} [ 911.288330] env[63355]: DEBUG oslo_vmware.api [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for the task: (returnval){ [ 911.288330] env[63355]: value = "task-1349694" [ 911.288330] env[63355]: _type = "Task" [ 911.288330] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.296090] env[63355]: DEBUG nova.scheduler.client.report [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Refreshing inventories for resource provider 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 911.301923] env[63355]: DEBUG oslo_vmware.api [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349694, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.309338] env[63355]: DEBUG nova.compute.manager [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 911.322096] env[63355]: DEBUG nova.scheduler.client.report [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Updating ProviderTree inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 911.322375] env[63355]: DEBUG nova.compute.provider_tree [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.332604] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349690, 'name': ReconfigVM_Task, 'duration_secs': 0.299991} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.334939] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.335326] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.335431] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.335604] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.335866] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.336030] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.336418] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.336686] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.336874] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.337132] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.337426] env[63355]: DEBUG nova.virt.hardware [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.338144] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 2539a79e-01c0-4e0c-aa66-8784441c6fda/2539a79e-01c0-4e0c-aa66-8784441c6fda.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.339389] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76634786-bdd3-4583-84d5-61e7fda43b3d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.342965] env[63355]: DEBUG nova.scheduler.client.report [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Refreshing aggregate associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, aggregates: None {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 911.345867] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1880e27-2a23-46d1-94f1-8e3cb1c7187b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.356095] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1f46a8-d2f2-4898-ba46-9efcc7eddf41 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.362165] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for the task: (returnval){ [ 911.362165] env[63355]: value = "task-1349695" [ 911.362165] env[63355]: _type = "Task" [ 911.362165] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.379080] env[63355]: DEBUG nova.scheduler.client.report [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Refreshing trait associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 911.391875] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349695, 'name': Rename_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.528261] env[63355]: DEBUG oslo_vmware.api [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349691, 'name': PowerOnVM_Task, 'duration_secs': 0.488544} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.528586] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.532292] env[63355]: DEBUG nova.compute.manager [None req-e5ba5452-c63d-469f-b9fc-e6353e5186aa tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.532780] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a5fc52-d4b5-42e6-a65b-277be102f368 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.541148] env[63355]: DEBUG nova.network.neutron [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 911.573977] env[63355]: DEBUG nova.compute.manager [req-e7c3b5c2-2eda-4623-ad76-9f2139907312 req-95f43c18-066a-47c3-85e8-96fa5ac66759 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Received event network-vif-plugged-62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.574400] env[63355]: DEBUG oslo_concurrency.lockutils [req-e7c3b5c2-2eda-4623-ad76-9f2139907312 req-95f43c18-066a-47c3-85e8-96fa5ac66759 service nova] Acquiring lock "74071e73-10fa-4dcb-aa15-91303b2278fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.576967] env[63355]: DEBUG oslo_concurrency.lockutils [req-e7c3b5c2-2eda-4623-ad76-9f2139907312 req-95f43c18-066a-47c3-85e8-96fa5ac66759 service nova] Lock "74071e73-10fa-4dcb-aa15-91303b2278fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.576967] env[63355]: DEBUG oslo_concurrency.lockutils [req-e7c3b5c2-2eda-4623-ad76-9f2139907312 req-95f43c18-066a-47c3-85e8-96fa5ac66759 service nova] Lock "74071e73-10fa-4dcb-aa15-91303b2278fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.576967] env[63355]: DEBUG nova.compute.manager [req-e7c3b5c2-2eda-4623-ad76-9f2139907312 req-95f43c18-066a-47c3-85e8-96fa5ac66759 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] No waiting events found dispatching network-vif-plugged-62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.576967] env[63355]: WARNING nova.compute.manager [req-e7c3b5c2-2eda-4623-ad76-9f2139907312 req-95f43c18-066a-47c3-85e8-96fa5ac66759 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Received unexpected event network-vif-plugged-62623c8d-eb2e-4e47-87f9-52f1df6a7d7f for instance with vm_state building and task_state spawning. [ 911.670382] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349692, 'name': CreateVM_Task, 'duration_secs': 0.318025} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.670522] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 911.670953] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.671137] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.671471] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 911.671846] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db1623f1-d2eb-4c11-8182-71389c9ac0b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.677785] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 911.677785] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523baa55-59ff-4357-a05f-944f458b7590" [ 911.677785] env[63355]: _type = "Task" [ 911.677785] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.681701] env[63355]: DEBUG nova.compute.manager [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.686113] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f9ab8f-5781-48cc-8f78-1be4f265fad8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.700337] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523baa55-59ff-4357-a05f-944f458b7590, 'name': SearchDatastore_Task, 'duration_secs': 0.009609} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.704774] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.705039] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.705974] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.705974] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.705974] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.706656] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc44b741-5803-4518-be82-a5ef693a3deb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.717060] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.717405] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.720709] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82aba9c9-e3d2-4a6f-a08e-64bd6cbd5db4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.730020] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 911.730020] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52898b03-e361-2d27-915f-5d477a8b9544" [ 911.730020] env[63355]: _type = "Task" [ 911.730020] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.738446] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52898b03-e361-2d27-915f-5d477a8b9544, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.750197] env[63355]: DEBUG nova.network.neutron [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Updating instance_info_cache with network_info: [{"id": "00f30816-4a44-44bc-9eb3-ca224f3da164", "address": "fa:16:3e:b9:7e:26", "network": {"id": "073b4af8-486c-4fa4-a17a-db867c0e1ff7", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-898609963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "214e0560beec42fea860b59ae67d374a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00f30816-4a", "ovs_interfaceid": "00f30816-4a44-44bc-9eb3-ca224f3da164", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.788185] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466deca1-2863-43d7-87b6-b0b1da03e26f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.795910] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9f3bf6-5ec9-4a0b-8400-1fd76a7dfa20 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.802394] env[63355]: DEBUG oslo_vmware.api [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Task: {'id': task-1349694, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169983} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.803138] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.803391] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.803667] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.803933] env[63355]: INFO nova.compute.manager [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Took 1.65 seconds to destroy the instance on the hypervisor. [ 911.805027] env[63355]: DEBUG oslo.service.loopingcall [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.805027] env[63355]: DEBUG nova.compute.manager [-] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.805027] env[63355]: DEBUG nova.network.neutron [-] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 911.833079] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3288e3-99aa-4ead-be10-edf4e8ca4af2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.843784] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64435dd6-0497-498b-8b1d-884ba57050cf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.863221] env[63355]: DEBUG nova.compute.provider_tree [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.874832] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349695, 'name': Rename_Task, 'duration_secs': 0.177473} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.876349] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 911.876908] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a95375f8-1fa4-4d35-8f42-fc315bea22ce {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.886228] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for the task: (returnval){ [ 911.886228] env[63355]: value = "task-1349696" [ 911.886228] env[63355]: _type = "Task" [ 911.886228] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.900142] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349696, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.181879] env[63355]: DEBUG nova.network.neutron [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Successfully updated port: 62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 912.206141] env[63355]: INFO nova.compute.manager [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] instance snapshotting [ 912.208872] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8032d1d-1bdf-4efa-9901-6641756f76b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.214098] env[63355]: DEBUG nova.compute.manager [req-d86aef30-c2a5-4292-a1db-38631c12fcf8 req-bdc9842d-e4ef-4deb-84e3-291ad91e776b service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Received event network-vif-deleted-fd4a5737-d1df-4e4e-aa70-3002a5778567 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.214289] env[63355]: INFO nova.compute.manager [req-d86aef30-c2a5-4292-a1db-38631c12fcf8 req-bdc9842d-e4ef-4deb-84e3-291ad91e776b service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Neutron deleted interface fd4a5737-d1df-4e4e-aa70-3002a5778567; detaching it from the instance and deleting it from the info cache [ 912.214466] env[63355]: DEBUG nova.network.neutron [req-d86aef30-c2a5-4292-a1db-38631c12fcf8 req-bdc9842d-e4ef-4deb-84e3-291ad91e776b service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.238461] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc64397f-7eb5-48e4-a8b3-db08b194514c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.253246] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52898b03-e361-2d27-915f-5d477a8b9544, 'name': SearchDatastore_Task, 'duration_secs': 0.011182} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.255030] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Releasing lock "refresh_cache-7c139710-d8d9-4cd7-bec0-6e021d3b2e68" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.255030] env[63355]: DEBUG nova.compute.manager [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Instance network_info: |[{"id": "00f30816-4a44-44bc-9eb3-ca224f3da164", "address": "fa:16:3e:b9:7e:26", "network": {"id": "073b4af8-486c-4fa4-a17a-db867c0e1ff7", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-898609963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "214e0560beec42fea860b59ae67d374a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00f30816-4a", "ovs_interfaceid": "00f30816-4a44-44bc-9eb3-ca224f3da164", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 912.255313] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-668dadd6-fe3b-45ad-af49-40b3464dbb32 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.257444] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:7e:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00f30816-4a44-44bc-9eb3-ca224f3da164', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.265246] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Creating folder: Project (214e0560beec42fea860b59ae67d374a). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 912.266084] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f196673-0ee7-4af3-90fa-c149a9523b8b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.274746] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 912.274746] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52837770-4eae-b447-5313-7a37fa305bc7" [ 912.274746] env[63355]: _type = "Task" [ 912.274746] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.282107] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Created folder: Project (214e0560beec42fea860b59ae67d374a) in parent group-v287607. [ 912.282107] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Creating folder: Instances. Parent ref: group-v287670. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 912.285700] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc04921b-eb23-4124-ae55-05875ef0ad5a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.288421] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52837770-4eae-b447-5313-7a37fa305bc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.303799] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Created folder: Instances in parent group-v287670. [ 912.304202] env[63355]: DEBUG oslo.service.loopingcall [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.304551] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 912.304887] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0707f5b-1da6-4edf-a150-c90a0e1ebbe9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.328290] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.328290] env[63355]: value = "task-1349699" [ 912.328290] env[63355]: _type = "Task" [ 912.328290] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.338626] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349699, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.366800] env[63355]: DEBUG nova.scheduler.client.report [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.400167] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349696, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.647283] env[63355]: DEBUG nova.network.neutron [-] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.686204] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.686204] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquired lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.686204] env[63355]: DEBUG nova.network.neutron [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.717334] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff6f44c8-5796-4429-909b-c2726facc0f9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.730069] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21f7acc-2778-4332-a93f-736028cc865e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.755613] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 912.764378] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ea1c673d-e175-4a3d-bc3d-dbf2d95b7054 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.767400] env[63355]: DEBUG nova.compute.manager [req-d86aef30-c2a5-4292-a1db-38631c12fcf8 req-bdc9842d-e4ef-4deb-84e3-291ad91e776b service nova] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Detach interface failed, port_id=fd4a5737-d1df-4e4e-aa70-3002a5778567, reason: Instance b6056441-9ee4-484f-a1d2-077546f2c581 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 912.776308] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 912.776308] env[63355]: value = "task-1349700" [ 912.776308] env[63355]: _type = "Task" [ 912.776308] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.792834] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52837770-4eae-b447-5313-7a37fa305bc7, 'name': SearchDatastore_Task, 'duration_secs': 0.017891} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.795525] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.796195] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] b4b09b1d-680e-47b8-aa8a-9b3d9167824d/b4b09b1d-680e-47b8-aa8a-9b3d9167824d.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.796430] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349700, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.796745] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ff50d4e-bb3c-433c-a808-4cb85e35758b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.807024] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 912.807024] env[63355]: value = "task-1349701" [ 912.807024] env[63355]: _type = "Task" [ 912.807024] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.822023] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.839797] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349699, 'name': CreateVM_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.876045] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.592s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.876045] env[63355]: DEBUG nova.compute.manager [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 912.879157] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.450s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.880368] env[63355]: INFO nova.compute.claims [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.899100] env[63355]: DEBUG oslo_vmware.api [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349696, 'name': PowerOnVM_Task, 'duration_secs': 0.544818} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.899408] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 912.899617] env[63355]: INFO nova.compute.manager [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Took 8.28 seconds to spawn the instance on the hypervisor. [ 912.899801] env[63355]: DEBUG nova.compute.manager [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.900622] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe84319c-d584-4170-a8d6-7f7463af4f1f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.150394] env[63355]: INFO nova.compute.manager [-] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Took 1.35 seconds to deallocate network for instance. [ 913.226505] env[63355]: DEBUG nova.network.neutron [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 913.293039] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349700, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.319163] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349701, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.345066] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349699, 'name': CreateVM_Task, 'duration_secs': 0.83146} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.345232] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 913.346160] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.346317] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.346787] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.347903] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3961c511-89e7-463d-a657-c1fc0f194a47 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.355022] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for the task: (returnval){ [ 913.355022] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52559db8-915c-f769-a735-dca84596116b" [ 913.355022] env[63355]: _type = "Task" [ 913.355022] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.365166] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52559db8-915c-f769-a735-dca84596116b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.385992] env[63355]: DEBUG nova.compute.utils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.391020] env[63355]: DEBUG nova.compute.manager [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 913.393842] env[63355]: DEBUG nova.network.neutron [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 913.431057] env[63355]: INFO nova.compute.manager [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Took 27.16 seconds to build instance. [ 913.435833] env[63355]: DEBUG nova.network.neutron [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Updating instance_info_cache with network_info: [{"id": "62623c8d-eb2e-4e47-87f9-52f1df6a7d7f", "address": "fa:16:3e:9e:33:fc", "network": {"id": "47dda8ee-ad14-4eba-99a4-7a3afd76415c", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-564920032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87a296f0ecd94879ab2336899fc575bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "edd47158-6f4b-44a1-8e82-0411205ad299", "external-id": "nsx-vlan-transportzone-587", "segmentation_id": 587, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62623c8d-eb", "ovs_interfaceid": "62623c8d-eb2e-4e47-87f9-52f1df6a7d7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.452762] env[63355]: DEBUG nova.policy [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54ad2682257b4250a8a96f4ba6a9523b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b3fee9bc99d49ea9de53d5dce52c79d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.473677] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "8e9ff9b7-636f-48df-9168-509d733278f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.474191] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e9ff9b7-636f-48df-9168-509d733278f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.658077] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.748964] env[63355]: DEBUG nova.network.neutron [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Successfully created port: 976ab0e0-5cea-4265-bc51-9b5cd29ed954 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 913.794078] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349700, 'name': CreateSnapshot_Task, 'duration_secs': 0.985787} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.794078] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 913.794789] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458cdf28-f0b5-4c5e-b865-878636db0a14 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.816864] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64851} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.817122] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] b4b09b1d-680e-47b8-aa8a-9b3d9167824d/b4b09b1d-680e-47b8-aa8a-9b3d9167824d.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 913.817365] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 913.817664] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea8aa44f-667e-4512-b318-4092ee49acdc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.827938] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 913.827938] env[63355]: value = "task-1349702" [ 913.827938] env[63355]: _type = "Task" [ 913.827938] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.839385] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349702, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.867021] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52559db8-915c-f769-a735-dca84596116b, 'name': SearchDatastore_Task, 'duration_secs': 0.055564} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.867021] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.867021] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.867021] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.867402] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.867402] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.867402] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-565ab41e-a56c-4b14-aebe-edf6e0c29cc0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.880189] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.880189] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 913.880189] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea4d1194-fc69-4b5d-8e1a-6d8c5bf9cf72 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.890201] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for the task: (returnval){ [ 913.890201] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52919c6b-a9b6-0229-e66b-739333ec5ff7" [ 913.890201] env[63355]: _type = "Task" [ 913.890201] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.893145] env[63355]: DEBUG nova.compute.manager [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 913.913145] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52919c6b-a9b6-0229-e66b-739333ec5ff7, 'name': SearchDatastore_Task, 'duration_secs': 0.018917} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.914750] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c985ce75-07cc-42fc-a614-e6724964d8dc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.923461] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for the task: (returnval){ [ 913.923461] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527bc7c1-6ddc-881e-2d8c-5364c44406b0" [ 913.923461] env[63355]: _type = "Task" [ 913.923461] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.935861] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f493de23-3ac9-42cf-afac-7ee90d79d941 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lock "2539a79e-01c0-4e0c-aa66-8784441c6fda" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.140s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.936427] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Releasing lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.936721] env[63355]: DEBUG nova.compute.manager [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Instance network_info: |[{"id": "62623c8d-eb2e-4e47-87f9-52f1df6a7d7f", "address": "fa:16:3e:9e:33:fc", "network": {"id": "47dda8ee-ad14-4eba-99a4-7a3afd76415c", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-564920032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87a296f0ecd94879ab2336899fc575bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "edd47158-6f4b-44a1-8e82-0411205ad299", "external-id": "nsx-vlan-transportzone-587", "segmentation_id": 587, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62623c8d-eb", "ovs_interfaceid": "62623c8d-eb2e-4e47-87f9-52f1df6a7d7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 913.937028] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527bc7c1-6ddc-881e-2d8c-5364c44406b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.941047] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:33:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'edd47158-6f4b-44a1-8e82-0411205ad299', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62623c8d-eb2e-4e47-87f9-52f1df6a7d7f', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.948906] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Creating folder: Project (87a296f0ecd94879ab2336899fc575bf). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.950355] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4e7940f-9eda-418d-92c2-6bf6a301ea38 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.965301] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Created folder: Project (87a296f0ecd94879ab2336899fc575bf) in parent group-v287607. [ 913.965643] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Creating folder: Instances. Parent ref: group-v287674. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.965941] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cb03a15d-8759-4220-91ef-8bc23f9de70e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.977396] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Created folder: Instances in parent group-v287674. [ 913.977685] env[63355]: DEBUG oslo.service.loopingcall [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.977943] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 913.979224] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3265a607-d8a0-4bd0-88a9-5bd267ff87cb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.008214] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.008214] env[63355]: value = "task-1349705" [ 914.008214] env[63355]: _type = "Task" [ 914.008214] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.021985] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349705, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.258124] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70329ae8-da48-4f24-bd11-34515e7e9356 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.269556] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfcd8353-1877-4404-9df4-3d3b71432712 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.302851] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e624a5-2c9c-4db8-9b90-f10c5ce9e6ca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.309497] env[63355]: DEBUG nova.compute.manager [req-3bdcbb68-6eeb-48a2-93d9-4b31c38ce280 req-43fb972f-017b-4fcb-9271-922d506b0c82 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Received event network-changed-62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.309497] env[63355]: DEBUG nova.compute.manager [req-3bdcbb68-6eeb-48a2-93d9-4b31c38ce280 req-43fb972f-017b-4fcb-9271-922d506b0c82 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Refreshing instance network info cache due to event network-changed-62623c8d-eb2e-4e47-87f9-52f1df6a7d7f. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 914.309497] env[63355]: DEBUG oslo_concurrency.lockutils [req-3bdcbb68-6eeb-48a2-93d9-4b31c38ce280 req-43fb972f-017b-4fcb-9271-922d506b0c82 service nova] Acquiring lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.309497] env[63355]: DEBUG oslo_concurrency.lockutils [req-3bdcbb68-6eeb-48a2-93d9-4b31c38ce280 req-43fb972f-017b-4fcb-9271-922d506b0c82 service nova] Acquired lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.309497] env[63355]: DEBUG nova.network.neutron [req-3bdcbb68-6eeb-48a2-93d9-4b31c38ce280 req-43fb972f-017b-4fcb-9271-922d506b0c82 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Refreshing network info cache for port 62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 914.319605] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 914.320448] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6479ebe2-fb0d-4f78-abc7-a1d1954d704b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.324466] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699ae296-b78c-4ba4-8dea-1c4a1d6653c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.347823] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349702, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092708} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.348507] env[63355]: DEBUG nova.compute.provider_tree [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.349648] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 914.349648] env[63355]: value = "task-1349706" [ 914.349648] env[63355]: _type = "Task" [ 914.349648] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.349894] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 914.350862] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066d0dbd-8983-4d8d-b7de-2633863adc05 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.364028] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349706, 'name': CloneVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.383439] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] b4b09b1d-680e-47b8-aa8a-9b3d9167824d/b4b09b1d-680e-47b8-aa8a-9b3d9167824d.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 914.386907] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a41bd15d-b82a-4eb8-ba49-15ee31f87f77 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.415344] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 914.415344] env[63355]: value = "task-1349707" [ 914.415344] env[63355]: _type = "Task" [ 914.415344] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.428118] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349707, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.438620] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527bc7c1-6ddc-881e-2d8c-5364c44406b0, 'name': SearchDatastore_Task, 'duration_secs': 0.018693} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.439031] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.439216] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 7c139710-d8d9-4cd7-bec0-6e021d3b2e68/7c139710-d8d9-4cd7-bec0-6e021d3b2e68.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 914.439489] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e270baba-9ba2-4650-a838-68a6bf446831 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.447364] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for the task: (returnval){ [ 914.447364] env[63355]: value = "task-1349708" [ 914.447364] env[63355]: _type = "Task" [ 914.447364] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.451341] env[63355]: DEBUG nova.compute.manager [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 914.460105] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349708, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.519442] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349705, 'name': CreateVM_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.858511] env[63355]: DEBUG nova.scheduler.client.report [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.869294] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349706, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.885962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquiring lock "2539a79e-01c0-4e0c-aa66-8784441c6fda" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.885962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lock "2539a79e-01c0-4e0c-aa66-8784441c6fda" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.885962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquiring lock "2539a79e-01c0-4e0c-aa66-8784441c6fda-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.885962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lock "2539a79e-01c0-4e0c-aa66-8784441c6fda-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.886137] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lock "2539a79e-01c0-4e0c-aa66-8784441c6fda-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.888292] env[63355]: INFO nova.compute.manager [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Terminating instance [ 914.890404] env[63355]: DEBUG nova.compute.manager [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.890614] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 914.891511] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22fd2e9-cf09-4648-8559-ba561142ad76 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.901709] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.901709] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63385c83-f38d-4d16-9cd9-ab1dcaa412d2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.911184] env[63355]: DEBUG nova.compute.manager [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.915080] env[63355]: DEBUG oslo_vmware.api [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for the task: (returnval){ [ 914.915080] env[63355]: value = "task-1349709" [ 914.915080] env[63355]: _type = "Task" [ 914.915080] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.936314] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.940740] env[63355]: DEBUG oslo_vmware.api [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349709, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.949653] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.949836] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.950034] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.950285] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.950494] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.950704] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.950955] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.951163] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.951354] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.951560] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.951783] env[63355]: DEBUG nova.virt.hardware [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.952862] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed9a633-3365-4f7f-a991-2b5f8d185f5e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.975563] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6615824c-c1c8-41d9-8792-e2791ca75292 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.980815] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349708, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.985483] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.022383] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349705, 'name': CreateVM_Task, 'duration_secs': 0.777297} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.022646] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.023522] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.023833] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.024303] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.025656] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0ccbe03-e1c0-4fac-a686-e7d23dd74105 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.033701] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 915.033701] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52746039-1333-d024-a77c-68e393baa8e2" [ 915.033701] env[63355]: _type = "Task" [ 915.033701] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.044826] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52746039-1333-d024-a77c-68e393baa8e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.143952] env[63355]: DEBUG nova.network.neutron [req-3bdcbb68-6eeb-48a2-93d9-4b31c38ce280 req-43fb972f-017b-4fcb-9271-922d506b0c82 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Updated VIF entry in instance network info cache for port 62623c8d-eb2e-4e47-87f9-52f1df6a7d7f. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 915.144421] env[63355]: DEBUG nova.network.neutron [req-3bdcbb68-6eeb-48a2-93d9-4b31c38ce280 req-43fb972f-017b-4fcb-9271-922d506b0c82 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Updating instance_info_cache with network_info: [{"id": "62623c8d-eb2e-4e47-87f9-52f1df6a7d7f", "address": "fa:16:3e:9e:33:fc", "network": {"id": "47dda8ee-ad14-4eba-99a4-7a3afd76415c", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-564920032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87a296f0ecd94879ab2336899fc575bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "edd47158-6f4b-44a1-8e82-0411205ad299", "external-id": "nsx-vlan-transportzone-587", "segmentation_id": 587, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62623c8d-eb", "ovs_interfaceid": "62623c8d-eb2e-4e47-87f9-52f1df6a7d7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.366794] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.366794] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.367970] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349706, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.368634] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.904s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.369999] env[63355]: INFO nova.compute.claims [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.412515] env[63355]: DEBUG nova.network.neutron [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Successfully updated port: 976ab0e0-5cea-4265-bc51-9b5cd29ed954 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.428817] env[63355]: DEBUG oslo_vmware.api [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349709, 'name': PowerOffVM_Task, 'duration_secs': 0.458642} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.429720] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 915.430160] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 915.430290] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55f36016-5130-4de0-9283-309a1ea0a6e0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.436054] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349707, 'name': ReconfigVM_Task, 'duration_secs': 0.651753} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.436746] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Reconfigured VM instance instance-00000042 to attach disk [datastore1] b4b09b1d-680e-47b8-aa8a-9b3d9167824d/b4b09b1d-680e-47b8-aa8a-9b3d9167824d.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.437808] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7fb5f7e2-8184-4c70-a4e0-fdb41a800769 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.445175] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 915.445175] env[63355]: value = "task-1349711" [ 915.445175] env[63355]: _type = "Task" [ 915.445175] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.459206] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349711, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.465219] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349708, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.724557} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.465516] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 7c139710-d8d9-4cd7-bec0-6e021d3b2e68/7c139710-d8d9-4cd7-bec0-6e021d3b2e68.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.465765] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.466083] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-265983f6-ad7c-434d-8c99-c764195638ac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.474874] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for the task: (returnval){ [ 915.474874] env[63355]: value = "task-1349712" [ 915.474874] env[63355]: _type = "Task" [ 915.474874] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.484336] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349712, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.512462] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 915.512747] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 915.512970] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Deleting the datastore file [datastore1] 2539a79e-01c0-4e0c-aa66-8784441c6fda {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.513242] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6f3e577-3ef4-4d21-85c3-510498a134fe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.522705] env[63355]: DEBUG oslo_vmware.api [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for the task: (returnval){ [ 915.522705] env[63355]: value = "task-1349713" [ 915.522705] env[63355]: _type = "Task" [ 915.522705] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.531888] env[63355]: DEBUG oslo_vmware.api [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349713, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.544198] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52746039-1333-d024-a77c-68e393baa8e2, 'name': SearchDatastore_Task, 'duration_secs': 0.059276} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.544510] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.544746] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.545080] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.545251] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.545436] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.545743] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7c37faa6-6b02-4aa9-9b4c-73aab584a26e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.555390] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.555643] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.556432] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad82f3b5-a7bc-4c3f-9a05-ff91a1b47a11 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.562513] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 915.562513] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c811d9-653c-7057-8b5c-18788ee0d910" [ 915.562513] env[63355]: _type = "Task" [ 915.562513] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.570886] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c811d9-653c-7057-8b5c-18788ee0d910, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.647224] env[63355]: DEBUG oslo_concurrency.lockutils [req-3bdcbb68-6eeb-48a2-93d9-4b31c38ce280 req-43fb972f-017b-4fcb-9271-922d506b0c82 service nova] Releasing lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.863510] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349706, 'name': CloneVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.869932] env[63355]: DEBUG nova.compute.utils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.871344] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.871528] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 915.920284] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.920284] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.920284] env[63355]: DEBUG nova.network.neutron [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.924751] env[63355]: DEBUG nova.policy [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4382845ae7694c018aa5c0562316583a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '092fc5912cd4449895f1c4d1318dd390', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.930310] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 915.930913] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 915.958939] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349711, 'name': Rename_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.985709] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349712, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072788} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.986075] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.986988] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ceb2cd-e256-4194-bbb5-97413e84c425 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.011571] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 7c139710-d8d9-4cd7-bec0-6e021d3b2e68/7c139710-d8d9-4cd7-bec0-6e021d3b2e68.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.011894] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbd5135e-a57b-4b52-b401-c642ebc145b1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.035827] env[63355]: DEBUG oslo_vmware.api [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Task: {'id': task-1349713, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179126} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.036116] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.036354] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 916.036511] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 916.036722] env[63355]: INFO nova.compute.manager [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Took 1.15 seconds to destroy the instance on the hypervisor. [ 916.037024] env[63355]: DEBUG oslo.service.loopingcall [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.038365] env[63355]: DEBUG nova.compute.manager [-] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 916.038473] env[63355]: DEBUG nova.network.neutron [-] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 916.040250] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for the task: (returnval){ [ 916.040250] env[63355]: value = "task-1349714" [ 916.040250] env[63355]: _type = "Task" [ 916.040250] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.049971] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349714, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.073327] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c811d9-653c-7057-8b5c-18788ee0d910, 'name': SearchDatastore_Task, 'duration_secs': 0.011656} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.074226] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c464369-437f-47d7-a7ec-2c8c57a76476 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.081866] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 916.081866] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e73c85-42a4-9187-fa89-7aea0595ab05" [ 916.081866] env[63355]: _type = "Task" [ 916.081866] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.090875] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e73c85-42a4-9187-fa89-7aea0595ab05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.336445] env[63355]: DEBUG nova.compute.manager [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Received event network-vif-plugged-976ab0e0-5cea-4265-bc51-9b5cd29ed954 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.336681] env[63355]: DEBUG oslo_concurrency.lockutils [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] Acquiring lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.336929] env[63355]: DEBUG oslo_concurrency.lockutils [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.337121] env[63355]: DEBUG oslo_concurrency.lockutils [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.337296] env[63355]: DEBUG nova.compute.manager [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] No waiting events found dispatching network-vif-plugged-976ab0e0-5cea-4265-bc51-9b5cd29ed954 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 916.337465] env[63355]: WARNING nova.compute.manager [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Received unexpected event network-vif-plugged-976ab0e0-5cea-4265-bc51-9b5cd29ed954 for instance with vm_state building and task_state spawning. [ 916.337628] env[63355]: DEBUG nova.compute.manager [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Received event network-changed-976ab0e0-5cea-4265-bc51-9b5cd29ed954 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.337783] env[63355]: DEBUG nova.compute.manager [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Refreshing instance network info cache due to event network-changed-976ab0e0-5cea-4265-bc51-9b5cd29ed954. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 916.337968] env[63355]: DEBUG oslo_concurrency.lockutils [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] Acquiring lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.348163] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Successfully created port: 42e50136-f618-4d11-82e9-5efca796fd5b {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.364290] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349706, 'name': CloneVM_Task, 'duration_secs': 1.604292} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.364565] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Created linked-clone VM from snapshot [ 916.365422] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b90c8e-8d6e-4246-89d4-eb6f8ef448fb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.373279] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Uploading image 98bf2e17-09e0-42e8-ab0d-b84472dc2ae4 {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 916.376659] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.407155] env[63355]: DEBUG oslo_vmware.rw_handles [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 916.407155] env[63355]: value = "vm-287677" [ 916.407155] env[63355]: _type = "VirtualMachine" [ 916.407155] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 916.407460] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-dc720c6f-5943-4cc3-8247-e8252e64b2e6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.416187] env[63355]: DEBUG oslo_vmware.rw_handles [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lease: (returnval){ [ 916.416187] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52abb867-4092-c8e7-6629-172349e84a8c" [ 916.416187] env[63355]: _type = "HttpNfcLease" [ 916.416187] env[63355]: } obtained for exporting VM: (result){ [ 916.416187] env[63355]: value = "vm-287677" [ 916.416187] env[63355]: _type = "VirtualMachine" [ 916.416187] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 916.416574] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the lease: (returnval){ [ 916.416574] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52abb867-4092-c8e7-6629-172349e84a8c" [ 916.416574] env[63355]: _type = "HttpNfcLease" [ 916.416574] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 916.426176] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.426176] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52abb867-4092-c8e7-6629-172349e84a8c" [ 916.426176] env[63355]: _type = "HttpNfcLease" [ 916.426176] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 916.447216] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 916.447475] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 916.447533] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Rebuilding the list of instances to heal {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 916.465895] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349711, 'name': Rename_Task, 'duration_secs': 0.8254} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.466458] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 916.466749] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d30cc0b8-56fe-43e4-97f0-8f7e58ffc080 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.476026] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 916.476026] env[63355]: value = "task-1349716" [ 916.476026] env[63355]: _type = "Task" [ 916.476026] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.487580] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349716, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.501374] env[63355]: DEBUG nova.network.neutron [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 916.553030] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349714, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.592713] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e73c85-42a4-9187-fa89-7aea0595ab05, 'name': SearchDatastore_Task, 'duration_secs': 0.011504} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.595434] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.595789] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 74071e73-10fa-4dcb-aa15-91303b2278fb/74071e73-10fa-4dcb-aa15-91303b2278fb.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 916.596282] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f82dff1-20f9-493f-8ae4-b5eed706b2b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.605127] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 916.605127] env[63355]: value = "task-1349717" [ 916.605127] env[63355]: _type = "Task" [ 916.605127] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.617686] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349717, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.725630] env[63355]: DEBUG nova.network.neutron [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance_info_cache with network_info: [{"id": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "address": "fa:16:3e:b3:26:57", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap976ab0e0-5c", "ovs_interfaceid": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.739454] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd4a0b0-9e86-4223-b332-d8691959b3a7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.750798] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffbfcf32-a634-46c3-9760-ada93939089e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.784424] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008ed0ee-cf4c-42ed-af5e-894cd3a5b736 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.793836] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21926c6-33dc-4397-b1de-4783876ce223 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.815163] env[63355]: DEBUG nova.compute.provider_tree [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.934380] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.934380] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52abb867-4092-c8e7-6629-172349e84a8c" [ 916.934380] env[63355]: _type = "HttpNfcLease" [ 916.934380] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 916.935489] env[63355]: DEBUG oslo_vmware.rw_handles [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 916.935489] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52abb867-4092-c8e7-6629-172349e84a8c" [ 916.935489] env[63355]: _type = "HttpNfcLease" [ 916.935489] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 916.936512] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3201d41-c00f-4318-b011-073dba145104 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.946222] env[63355]: DEBUG oslo_vmware.rw_handles [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c3cc-d8d0-75ae-83bf-d0e0b7dc574e/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 916.946451] env[63355]: DEBUG oslo_vmware.rw_handles [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c3cc-d8d0-75ae-83bf-d0e0b7dc574e/disk-0.vmdk for reading. {{(pid=63355) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 917.010212] env[63355]: DEBUG nova.network.neutron [-] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.016057] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Skipping network cache update for instance because it is being deleted. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 917.016262] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 917.016395] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 917.016608] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 917.016818] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 917.017064] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 917.038114] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349716, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.059267] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349714, 'name': ReconfigVM_Task, 'duration_secs': 0.769252} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.060433] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 7c139710-d8d9-4cd7-bec0-6e021d3b2e68/7c139710-d8d9-4cd7-bec0-6e021d3b2e68.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.061565] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ab3817b-9e43-4ed2-a813-1ee9885788c7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.073229] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for the task: (returnval){ [ 917.073229] env[63355]: value = "task-1349718" [ 917.073229] env[63355]: _type = "Task" [ 917.073229] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.085950] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349718, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.089392] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.089553] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquired lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.089714] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Forcefully refreshing network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 917.089962] env[63355]: DEBUG nova.objects.instance [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lazy-loading 'info_cache' on Instance uuid 386e847e-967b-4247-9730-cdc5ac251474 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.110112] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f8b70032-bbf6-47e3-a4d9-426143cbaa84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.122774] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349717, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.228506] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.228934] env[63355]: DEBUG nova.compute.manager [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Instance network_info: |[{"id": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "address": "fa:16:3e:b3:26:57", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap976ab0e0-5c", "ovs_interfaceid": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 917.229231] env[63355]: DEBUG oslo_concurrency.lockutils [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] Acquired lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.229397] env[63355]: DEBUG nova.network.neutron [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Refreshing network info cache for port 976ab0e0-5cea-4265-bc51-9b5cd29ed954 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 917.230734] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:26:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '976ab0e0-5cea-4265-bc51-9b5cd29ed954', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.239532] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating folder: Project (6b3fee9bc99d49ea9de53d5dce52c79d). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.240185] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15f60f7b-4312-4b70-9504-83f38d452e69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.255966] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Created folder: Project (6b3fee9bc99d49ea9de53d5dce52c79d) in parent group-v287607. [ 917.256167] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating folder: Instances. Parent ref: group-v287678. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.256440] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffb9b17a-4b01-4999-af4a-d1d965a840f3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.269997] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Created folder: Instances in parent group-v287678. [ 917.270322] env[63355]: DEBUG oslo.service.loopingcall [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.270510] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.270735] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d7d63f83-fc8f-43f8-a4bf-e2aba8b2a4a9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.292640] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.292640] env[63355]: value = "task-1349721" [ 917.292640] env[63355]: _type = "Task" [ 917.292640] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.304801] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349721, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.318562] env[63355]: DEBUG nova.scheduler.client.report [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.389296] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.413314] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.413630] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.413782] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.413962] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.414128] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.414326] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.414497] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.414656] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.414825] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.415016] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.415243] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.416216] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c8ba3a-2541-4ac8-9b3f-a4b761c6d701 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.424944] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911b1b7a-6c1d-47d7-83ca-45e80c6406f2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.513636] env[63355]: INFO nova.compute.manager [-] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Took 1.47 seconds to deallocate network for instance. [ 917.530828] env[63355]: DEBUG oslo_vmware.api [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349716, 'name': PowerOnVM_Task, 'duration_secs': 0.655619} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.531346] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 917.531754] env[63355]: DEBUG nova.compute.manager [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.532686] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19bad9c3-03f7-4ec2-abb3-0d998c23bde6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.584321] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349718, 'name': Rename_Task, 'duration_secs': 0.235254} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.586050] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.586050] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cdd70009-15f6-4bd1-9e3e-5333e177ed26 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.599026] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for the task: (returnval){ [ 917.599026] env[63355]: value = "task-1349722" [ 917.599026] env[63355]: _type = "Task" [ 917.599026] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.613974] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349722, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.626660] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349717, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.755539} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.630026] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 74071e73-10fa-4dcb-aa15-91303b2278fb/74071e73-10fa-4dcb-aa15-91303b2278fb.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 917.630026] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.630026] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fdddede-c7f7-4ce8-bd51-992b7d47356b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.637299] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 917.637299] env[63355]: value = "task-1349723" [ 917.637299] env[63355]: _type = "Task" [ 917.637299] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.649999] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349723, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.807517] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349721, 'name': CreateVM_Task, 'duration_secs': 0.383025} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.807716] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 917.808645] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.809070] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.811661] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 917.811661] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-519f00e5-1d91-40d5-be05-17012c67e0e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.816231] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 917.816231] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525f62f0-1278-2d79-4317-f6ee003c6604" [ 917.816231] env[63355]: _type = "Task" [ 917.816231] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.827270] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.827852] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.831067] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525f62f0-1278-2d79-4317-f6ee003c6604, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.831067] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.666s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.836709] env[63355]: INFO nova.compute.claims [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.989966] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Successfully updated port: 42e50136-f618-4d11-82e9-5efca796fd5b {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.022197] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.057462] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.116256] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349722, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.150684] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349723, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084034} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.150684] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.150684] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b67b032-6f6a-477b-9821-65fe02d0660d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.178787] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 74071e73-10fa-4dcb-aa15-91303b2278fb/74071e73-10fa-4dcb-aa15-91303b2278fb.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.180724] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89ad70a6-b09a-4a8e-8e71-8f3b9d3fe678 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.205915] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 918.205915] env[63355]: value = "task-1349724" [ 918.205915] env[63355]: _type = "Task" [ 918.205915] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.217909] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349724, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.294135] env[63355]: DEBUG nova.network.neutron [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updated VIF entry in instance network info cache for port 976ab0e0-5cea-4265-bc51-9b5cd29ed954. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 918.294656] env[63355]: DEBUG nova.network.neutron [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance_info_cache with network_info: [{"id": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "address": "fa:16:3e:b3:26:57", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap976ab0e0-5c", "ovs_interfaceid": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.328669] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525f62f0-1278-2d79-4317-f6ee003c6604, 'name': SearchDatastore_Task, 'duration_secs': 0.018374} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.329276] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.329651] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.329968] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.330224] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.330472] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.330851] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-797ac677-1bea-4c49-b9f4-b916e2756f23 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.343268] env[63355]: DEBUG nova.compute.utils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.347225] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.347496] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 918.351756] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.352408] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 918.354514] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e09742f3-328e-464c-90da-82986cf3d913 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.365270] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 918.365270] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ead9d5-6926-229e-0735-290f52594885" [ 918.365270] env[63355]: _type = "Task" [ 918.365270] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.378423] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ead9d5-6926-229e-0735-290f52594885, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.426780] env[63355]: DEBUG nova.policy [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4382845ae7694c018aa5c0562316583a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '092fc5912cd4449895f1c4d1318dd390', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 918.496042] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "refresh_cache-dc56b350-cee7-49c4-9712-8c8c29146ff2" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.496042] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired lock "refresh_cache-dc56b350-cee7-49c4-9712-8c8c29146ff2" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.496042] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.611324] env[63355]: DEBUG oslo_vmware.api [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349722, 'name': PowerOnVM_Task, 'duration_secs': 0.787724} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.612140] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.615515] env[63355]: INFO nova.compute.manager [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Took 9.71 seconds to spawn the instance on the hypervisor. [ 918.615515] env[63355]: DEBUG nova.compute.manager [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.615515] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b84cd3f-d8bb-4667-801a-d58dd4ce60c0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.719605] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349724, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.729522] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Successfully created port: b6165b88-87f9-4c10-bc5c-fc58b48d167c {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.797502] env[63355]: DEBUG oslo_concurrency.lockutils [req-9ebfecb1-35ba-45a3-8cae-526ec7849e8b req-6f751347-ed16-460a-865c-950c7b8721f1 service nova] Releasing lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.847070] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.872863] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [{"id": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "address": "fa:16:3e:a9:13:18", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348dcf0-b9", "ovs_interfaceid": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.883661] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ead9d5-6926-229e-0735-290f52594885, 'name': SearchDatastore_Task, 'duration_secs': 0.017536} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.885411] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4cf4596-3df8-4664-b02c-44689406d277 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.899941] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 918.899941] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523742d1-55c6-9bfb-afb2-536e3526cb1c" [ 918.899941] env[63355]: _type = "Task" [ 918.899941] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.912785] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523742d1-55c6-9bfb-afb2-536e3526cb1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.036100] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 919.138408] env[63355]: INFO nova.compute.manager [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Took 29.46 seconds to build instance. [ 919.188784] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e3ba7ed-05af-4fbf-a396-8e1b68afd7b4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.193497] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Updating instance_info_cache with network_info: [{"id": "42e50136-f618-4d11-82e9-5efca796fd5b", "address": "fa:16:3e:2f:7d:cf", "network": {"id": "e16cdd31-8271-4c70-90fd-3943b04a95c9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-902402066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "092fc5912cd4449895f1c4d1318dd390", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e50136-f6", "ovs_interfaceid": "42e50136-f618-4d11-82e9-5efca796fd5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.201601] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aeaad4a-bd32-43f5-a9d2-f3bb3dc7922c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.240171] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622304c0-77a8-4590-a0fb-af42111448b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.246305] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349724, 'name': ReconfigVM_Task, 'duration_secs': 0.524795} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.247017] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 74071e73-10fa-4dcb-aa15-91303b2278fb/74071e73-10fa-4dcb-aa15-91303b2278fb.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.247686] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd4d1d50-fae5-4d72-99b6-aaf8a6771f1b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.252858] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a4965c-760c-40ed-afa6-5411e9d88a29 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.258863] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 919.258863] env[63355]: value = "task-1349725" [ 919.258863] env[63355]: _type = "Task" [ 919.258863] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.270780] env[63355]: DEBUG nova.compute.provider_tree [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.278680] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349725, 'name': Rename_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.379611] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Releasing lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.380237] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updated the network info_cache for instance {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 919.380237] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 919.380553] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 919.380814] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 919.381095] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 919.381311] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 919.381512] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 919.381645] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 919.382295] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 919.414216] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523742d1-55c6-9bfb-afb2-536e3526cb1c, 'name': SearchDatastore_Task, 'duration_secs': 0.019183} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.414501] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.414926] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5/537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.415247] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec0d1faa-2e21-4253-a573-ee60aea10cfe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.425752] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 919.425752] env[63355]: value = "task-1349726" [ 919.425752] env[63355]: _type = "Task" [ 919.425752] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.435432] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349726, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.640406] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b6c1689-67e5-4123-a349-0b1251528984 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.524s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.697412] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Releasing lock "refresh_cache-dc56b350-cee7-49c4-9712-8c8c29146ff2" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.697757] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Instance network_info: |[{"id": "42e50136-f618-4d11-82e9-5efca796fd5b", "address": "fa:16:3e:2f:7d:cf", "network": {"id": "e16cdd31-8271-4c70-90fd-3943b04a95c9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-902402066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "092fc5912cd4449895f1c4d1318dd390", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e50136-f6", "ovs_interfaceid": "42e50136-f618-4d11-82e9-5efca796fd5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 919.698258] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:7d:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a55f45a-d631-4ebc-b73b-8a30bd0a32a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '42e50136-f618-4d11-82e9-5efca796fd5b', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.707476] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Creating folder: Project (092fc5912cd4449895f1c4d1318dd390). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 919.707986] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-60b46744-17e3-4ab9-bd5b-6e2edb8b767a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.723941] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Created folder: Project (092fc5912cd4449895f1c4d1318dd390) in parent group-v287607. [ 919.725176] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Creating folder: Instances. Parent ref: group-v287681. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 919.726037] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d020ab6c-ac24-4ae1-85f9-4a8f26721559 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.742060] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Created folder: Instances in parent group-v287681. [ 919.742414] env[63355]: DEBUG oslo.service.loopingcall [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.742667] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 919.743013] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f335d252-eb48-409d-a8d7-99a6a2e09940 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.766236] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.766236] env[63355]: value = "task-1349729" [ 919.766236] env[63355]: _type = "Task" [ 919.766236] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.776386] env[63355]: DEBUG oslo_vmware.rw_handles [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521e74fe-5003-7980-15f1-e31689e12e7b/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 919.777411] env[63355]: DEBUG nova.scheduler.client.report [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.781315] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cedf7ad-61d3-4e0f-b3f3-1a3af4755d36 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.788271] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349725, 'name': Rename_Task, 'duration_secs': 0.195206} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.789010] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.789315] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac5b66eb-2690-4980-a7be-bc01b4591cf5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.796846] env[63355]: DEBUG oslo_vmware.rw_handles [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521e74fe-5003-7980-15f1-e31689e12e7b/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 919.797050] env[63355]: ERROR oslo_vmware.rw_handles [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521e74fe-5003-7980-15f1-e31689e12e7b/disk-0.vmdk due to incomplete transfer. [ 919.797327] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349729, 'name': CreateVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.798595] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-90b317d1-6738-4550-bf89-8ed1ec18abab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.805192] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 919.805192] env[63355]: value = "task-1349730" [ 919.805192] env[63355]: _type = "Task" [ 919.805192] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.810562] env[63355]: DEBUG oslo_vmware.rw_handles [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521e74fe-5003-7980-15f1-e31689e12e7b/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 919.810795] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Uploaded image f755e516-c1d3-4618-aedf-9caa138936ff to the Glance image server {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 919.813346] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 919.814203] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ba7ece04-71e8-42fe-bc8f-abb635880d9e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.821325] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349730, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.823529] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 919.823529] env[63355]: value = "task-1349731" [ 919.823529] env[63355]: _type = "Task" [ 919.823529] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.833233] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349731, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.860800] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.886286] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.886594] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.886878] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.887154] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.887401] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.887674] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.887904] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.888130] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.888355] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.888575] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.888803] env[63355]: DEBUG nova.virt.hardware [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.889991] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.890969] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855f0f25-af24-4bca-b13e-fc4259bbb069 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.901232] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c1aaa7-12f6-4c2c-806c-2aca887b69cf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.938108] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349726, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.965938] env[63355]: DEBUG nova.compute.manager [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Received event network-vif-deleted-31d968dc-6934-46fe-86bd-fab6751b337c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.966241] env[63355]: DEBUG nova.compute.manager [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Received event network-vif-plugged-42e50136-f618-4d11-82e9-5efca796fd5b {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.966476] env[63355]: DEBUG oslo_concurrency.lockutils [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] Acquiring lock "dc56b350-cee7-49c4-9712-8c8c29146ff2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.966737] env[63355]: DEBUG oslo_concurrency.lockutils [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] Lock "dc56b350-cee7-49c4-9712-8c8c29146ff2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.966944] env[63355]: DEBUG oslo_concurrency.lockutils [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] Lock "dc56b350-cee7-49c4-9712-8c8c29146ff2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.967145] env[63355]: DEBUG nova.compute.manager [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] No waiting events found dispatching network-vif-plugged-42e50136-f618-4d11-82e9-5efca796fd5b {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 919.967326] env[63355]: WARNING nova.compute.manager [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Received unexpected event network-vif-plugged-42e50136-f618-4d11-82e9-5efca796fd5b for instance with vm_state building and task_state spawning. [ 919.967532] env[63355]: DEBUG nova.compute.manager [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Received event network-changed-42e50136-f618-4d11-82e9-5efca796fd5b {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.967725] env[63355]: DEBUG nova.compute.manager [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Refreshing instance network info cache due to event network-changed-42e50136-f618-4d11-82e9-5efca796fd5b. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 919.967977] env[63355]: DEBUG oslo_concurrency.lockutils [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] Acquiring lock "refresh_cache-dc56b350-cee7-49c4-9712-8c8c29146ff2" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.968172] env[63355]: DEBUG oslo_concurrency.lockutils [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] Acquired lock "refresh_cache-dc56b350-cee7-49c4-9712-8c8c29146ff2" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.968375] env[63355]: DEBUG nova.network.neutron [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Refreshing network info cache for port 42e50136-f618-4d11-82e9-5efca796fd5b {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 920.143236] env[63355]: DEBUG nova.compute.manager [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 920.279052] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349729, 'name': CreateVM_Task, 'duration_secs': 0.420592} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.279052] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 920.279052] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.279052] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.279052] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 920.279538] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b1bb647-2cb3-4075-9514-57108b4c12fe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.285241] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.285857] env[63355]: DEBUG nova.compute.manager [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 920.290499] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.820s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.292299] env[63355]: INFO nova.compute.claims [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.295461] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 920.295461] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c4c4ec-e6a2-52cd-6986-36ba1fa90730" [ 920.295461] env[63355]: _type = "Task" [ 920.295461] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.308836] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c4c4ec-e6a2-52cd-6986-36ba1fa90730, 'name': SearchDatastore_Task, 'duration_secs': 0.016802} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.313240] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.314344] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.314750] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.315137] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.315710] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.318807] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b013854-d811-4bb5-9f2c-c9c607c6bea1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.327284] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349730, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.333414] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.333885] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 920.334871] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-480ca1ec-f324-4a46-afd1-8a47c8417324 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.345688] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 920.345688] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526274ec-ec5f-5c5e-5551-33f1c4a006f2" [ 920.345688] env[63355]: _type = "Task" [ 920.345688] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.349472] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349731, 'name': Destroy_Task} progress is 33%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.359988] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526274ec-ec5f-5c5e-5551-33f1c4a006f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.441024] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349726, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.600909} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.441024] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5/537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 920.441024] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.441024] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f715d13-0550-4358-9c32-cf2a41fc71eb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.448215] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 920.448215] env[63355]: value = "task-1349732" [ 920.448215] env[63355]: _type = "Task" [ 920.448215] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.458335] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349732, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.575376] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Successfully updated port: b6165b88-87f9-4c10-bc5c-fc58b48d167c {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 920.623023] env[63355]: DEBUG nova.compute.manager [req-ae73def2-502d-4669-a2c0-985ba25b6487 req-10e8bf43-1ae1-4a8d-b36b-e6ae2ba44133 service nova] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Received event network-vif-plugged-b6165b88-87f9-4c10-bc5c-fc58b48d167c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.623023] env[63355]: DEBUG oslo_concurrency.lockutils [req-ae73def2-502d-4669-a2c0-985ba25b6487 req-10e8bf43-1ae1-4a8d-b36b-e6ae2ba44133 service nova] Acquiring lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.623504] env[63355]: DEBUG oslo_concurrency.lockutils [req-ae73def2-502d-4669-a2c0-985ba25b6487 req-10e8bf43-1ae1-4a8d-b36b-e6ae2ba44133 service nova] Lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.623504] env[63355]: DEBUG oslo_concurrency.lockutils [req-ae73def2-502d-4669-a2c0-985ba25b6487 req-10e8bf43-1ae1-4a8d-b36b-e6ae2ba44133 service nova] Lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.623681] env[63355]: DEBUG nova.compute.manager [req-ae73def2-502d-4669-a2c0-985ba25b6487 req-10e8bf43-1ae1-4a8d-b36b-e6ae2ba44133 service nova] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] No waiting events found dispatching network-vif-plugged-b6165b88-87f9-4c10-bc5c-fc58b48d167c {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.623875] env[63355]: WARNING nova.compute.manager [req-ae73def2-502d-4669-a2c0-985ba25b6487 req-10e8bf43-1ae1-4a8d-b36b-e6ae2ba44133 service nova] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Received unexpected event network-vif-plugged-b6165b88-87f9-4c10-bc5c-fc58b48d167c for instance with vm_state building and task_state spawning. [ 920.669142] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.693256] env[63355]: DEBUG nova.network.neutron [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Updated VIF entry in instance network info cache for port 42e50136-f618-4d11-82e9-5efca796fd5b. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 920.693693] env[63355]: DEBUG nova.network.neutron [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Updating instance_info_cache with network_info: [{"id": "42e50136-f618-4d11-82e9-5efca796fd5b", "address": "fa:16:3e:2f:7d:cf", "network": {"id": "e16cdd31-8271-4c70-90fd-3943b04a95c9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-902402066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "092fc5912cd4449895f1c4d1318dd390", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e50136-f6", "ovs_interfaceid": "42e50136-f618-4d11-82e9-5efca796fd5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.801606] env[63355]: DEBUG nova.compute.utils [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.805023] env[63355]: DEBUG nova.compute.manager [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 920.805023] env[63355]: DEBUG nova.network.neutron [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 920.824299] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349730, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.839871] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349731, 'name': Destroy_Task} progress is 33%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.852254] env[63355]: DEBUG nova.policy [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6da01f502b6b4e3db9c29aeba89adc6e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c787ea7464c54aa9a443b4945e8cb5df', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.866071] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526274ec-ec5f-5c5e-5551-33f1c4a006f2, 'name': SearchDatastore_Task, 'duration_secs': 0.017129} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.866800] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9830a788-ad33-4fff-ab28-dc7193be622f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.873099] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 920.873099] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a34339-aae1-8883-4ec8-5f7282f6175f" [ 920.873099] env[63355]: _type = "Task" [ 920.873099] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.882205] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a34339-aae1-8883-4ec8-5f7282f6175f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.961242] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349732, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.358462} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.961547] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.962357] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a18a4d-8bfc-479a-940c-dd0b43a6dc84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.987897] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5/537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.988310] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c08e39e1-0d65-4dee-b563-5b9a91aba634 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.015718] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 921.015718] env[63355]: value = "task-1349733" [ 921.015718] env[63355]: _type = "Task" [ 921.015718] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.025531] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349733, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.077872] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "refresh_cache-199bc488-2e5d-4cea-aefb-ddd35ecc7a30" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.078135] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired lock "refresh_cache-199bc488-2e5d-4cea-aefb-ddd35ecc7a30" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.078330] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 921.136504] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "b4b09b1d-680e-47b8-aa8a-9b3d9167824d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.136504] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "b4b09b1d-680e-47b8-aa8a-9b3d9167824d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.136504] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "b4b09b1d-680e-47b8-aa8a-9b3d9167824d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.136504] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "b4b09b1d-680e-47b8-aa8a-9b3d9167824d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.136717] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "b4b09b1d-680e-47b8-aa8a-9b3d9167824d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.138579] env[63355]: INFO nova.compute.manager [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Terminating instance [ 921.140342] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "refresh_cache-b4b09b1d-680e-47b8-aa8a-9b3d9167824d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.140592] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquired lock "refresh_cache-b4b09b1d-680e-47b8-aa8a-9b3d9167824d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.140722] env[63355]: DEBUG nova.network.neutron [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 921.188828] env[63355]: DEBUG nova.network.neutron [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Successfully created port: 0500cf25-c654-4748-a58a-8e51bd919768 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.196430] env[63355]: DEBUG oslo_concurrency.lockutils [req-29d5215a-5de0-4c09-9ee7-fbae4e2cefb6 req-fa4c058d-98f7-49a4-a275-704a513ac9ce service nova] Releasing lock "refresh_cache-dc56b350-cee7-49c4-9712-8c8c29146ff2" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.282340] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquiring lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.282738] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.283084] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquiring lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.283295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.283474] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.286640] env[63355]: INFO nova.compute.manager [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Terminating instance [ 921.289378] env[63355]: DEBUG nova.compute.manager [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 921.289588] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 921.290454] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73782b35-37c5-4810-adb9-c37d3b4d66ec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.299732] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 921.300044] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fcebb9ac-d988-4084-8193-678d41e6cfe9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.308659] env[63355]: DEBUG nova.compute.manager [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 921.313908] env[63355]: DEBUG oslo_vmware.api [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for the task: (returnval){ [ 921.313908] env[63355]: value = "task-1349734" [ 921.313908] env[63355]: _type = "Task" [ 921.313908] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.331236] env[63355]: DEBUG oslo_vmware.api [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349734, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.331557] env[63355]: DEBUG oslo_vmware.api [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349730, 'name': PowerOnVM_Task, 'duration_secs': 1.063851} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.339523] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.339523] env[63355]: INFO nova.compute.manager [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Took 10.03 seconds to spawn the instance on the hypervisor. [ 921.339523] env[63355]: DEBUG nova.compute.manager [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.339523] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27f8c18-e50b-4c5d-bdf5-81f751a13492 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.347202] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349731, 'name': Destroy_Task, 'duration_secs': 1.113372} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.349209] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Destroyed the VM [ 921.349487] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 921.352397] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1d7fd231-d39c-4616-9da5-d7c82ab1d077 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.360515] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 921.360515] env[63355]: value = "task-1349735" [ 921.360515] env[63355]: _type = "Task" [ 921.360515] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.375793] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349735, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.386867] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a34339-aae1-8883-4ec8-5f7282f6175f, 'name': SearchDatastore_Task, 'duration_secs': 0.014739} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.390041] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.390360] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] dc56b350-cee7-49c4-9712-8c8c29146ff2/dc56b350-cee7-49c4-9712-8c8c29146ff2.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.390825] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-048b3bf5-a520-4dd4-83d2-389d31776440 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.400196] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 921.400196] env[63355]: value = "task-1349736" [ 921.400196] env[63355]: _type = "Task" [ 921.400196] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.413200] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349736, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.526363] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349733, 'name': ReconfigVM_Task, 'duration_secs': 0.333567} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.529711] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5/537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.530620] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9653ca2d-502c-48c5-b2e8-f6bf61bdb947 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.540643] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 921.540643] env[63355]: value = "task-1349737" [ 921.540643] env[63355]: _type = "Task" [ 921.540643] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.562189] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349737, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.631025] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.653483] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "248ec2f2-3a41-444a-b550-71c5438a45f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.653901] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "248ec2f2-3a41-444a-b550-71c5438a45f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.672786] env[63355]: DEBUG nova.network.neutron [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.688617] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e35a75e-79e0-473c-81f2-17fa935aa058 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.704937] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ef3341-156a-440a-b900-f5eb375904f7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.750055] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a56fbe-6e71-43fe-8229-cd8174561a89 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.756763] env[63355]: DEBUG nova.network.neutron [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.762283] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07dbf860-2c0e-4541-ad40-008813076ec6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.779211] env[63355]: DEBUG nova.compute.provider_tree [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 921.816118] env[63355]: INFO nova.virt.block_device [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Booting with volume 16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1 at /dev/sda [ 921.838869] env[63355]: DEBUG oslo_vmware.api [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349734, 'name': PowerOffVM_Task, 'duration_secs': 0.300084} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.839196] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.839375] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.839718] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6cc1556c-42bc-450a-9e1a-9cd6f66b54ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.869524] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8020975e-4a16-47d7-9fdf-f10dc186c8bf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.875815] env[63355]: INFO nova.compute.manager [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Took 30.64 seconds to build instance. [ 921.883564] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349735, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.888016] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c1a375-3388-4115-88f5-cc4b8f239b0d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.914190] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349736, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.927842] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-28fef045-d577-42ef-b274-8c6d40e1f3f9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.930103] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.930461] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.930703] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Deleting the datastore file [datastore1] 7c139710-d8d9-4cd7-bec0-6e021d3b2e68 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.935404] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03adc9ae-e29c-4bb0-808e-eff2bf4e47e7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.948718] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3022a79c-ab6d-4885-94ca-014da5d9db1e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.962098] env[63355]: DEBUG oslo_vmware.api [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for the task: (returnval){ [ 921.962098] env[63355]: value = "task-1349739" [ 921.962098] env[63355]: _type = "Task" [ 921.962098] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.973446] env[63355]: DEBUG oslo_vmware.api [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349739, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.992483] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6905c2c-8a38-4606-b16b-425d4910eab5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.005456] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4336c6-7ed0-43c3-a08e-de439a69d6e7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.025222] env[63355]: DEBUG nova.virt.block_device [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Updating existing volume attachment record: d2995d0e-77cd-4387-be67-509b1ccf3a99 {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 922.030595] env[63355]: DEBUG nova.network.neutron [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Updating instance_info_cache with network_info: [{"id": "b6165b88-87f9-4c10-bc5c-fc58b48d167c", "address": "fa:16:3e:b0:3a:09", "network": {"id": "e16cdd31-8271-4c70-90fd-3943b04a95c9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-902402066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "092fc5912cd4449895f1c4d1318dd390", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6165b88-87", "ovs_interfaceid": "b6165b88-87f9-4c10-bc5c-fc58b48d167c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.055109] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349737, 'name': Rename_Task, 'duration_secs': 0.166665} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.055445] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.055756] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b521c9ec-3f24-4579-993c-5a195ec98c7f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.065036] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 922.065036] env[63355]: value = "task-1349740" [ 922.065036] env[63355]: _type = "Task" [ 922.065036] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.075182] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349740, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.261858] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Releasing lock "refresh_cache-b4b09b1d-680e-47b8-aa8a-9b3d9167824d" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.262548] env[63355]: DEBUG nova.compute.manager [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.262685] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 922.263751] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8a1a50-37f2-42b0-a2e0-175bb4016801 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.275189] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.275510] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1fe142c8-5846-401e-b7c1-cdef5a7a0de8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.288587] env[63355]: DEBUG oslo_vmware.api [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 922.288587] env[63355]: value = "task-1349741" [ 922.288587] env[63355]: _type = "Task" [ 922.288587] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.298386] env[63355]: DEBUG oslo_vmware.api [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349741, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.305348] env[63355]: ERROR nova.scheduler.client.report [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [req-99ab4f18-dd4a-46f2-baf9-74284845a12c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 47abb610-db7e-4770-911d-187dd075ef8b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-99ab4f18-dd4a-46f2-baf9-74284845a12c"}]} [ 922.326880] env[63355]: DEBUG nova.scheduler.client.report [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Refreshing inventories for resource provider 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 922.344539] env[63355]: DEBUG nova.scheduler.client.report [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Updating ProviderTree inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 922.344899] env[63355]: DEBUG nova.compute.provider_tree [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 922.357352] env[63355]: DEBUG nova.scheduler.client.report [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Refreshing aggregate associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, aggregates: None {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 922.374038] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349735, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.378723] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12833639-9a45-4a4e-8b47-cfef4f15a3a4 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "74071e73-10fa-4dcb-aa15-91303b2278fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.101s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.379841] env[63355]: DEBUG nova.scheduler.client.report [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Refreshing trait associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 922.416283] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349736, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.624317} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.416581] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] dc56b350-cee7-49c4-9712-8c8c29146ff2/dc56b350-cee7-49c4-9712-8c8c29146ff2.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.416805] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.417083] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7d523357-74cb-4cf7-a145-1cbed1b2a383 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.426904] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 922.426904] env[63355]: value = "task-1349742" [ 922.426904] env[63355]: _type = "Task" [ 922.426904] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.439421] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349742, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.477370] env[63355]: DEBUG oslo_vmware.api [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Task: {'id': task-1349739, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249609} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.477657] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.477858] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 922.478093] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 922.478257] env[63355]: INFO nova.compute.manager [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Took 1.19 seconds to destroy the instance on the hypervisor. [ 922.478508] env[63355]: DEBUG oslo.service.loopingcall [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.478748] env[63355]: DEBUG nova.compute.manager [-] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.478832] env[63355]: DEBUG nova.network.neutron [-] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 922.534077] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Releasing lock "refresh_cache-199bc488-2e5d-4cea-aefb-ddd35ecc7a30" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.534415] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Instance network_info: |[{"id": "b6165b88-87f9-4c10-bc5c-fc58b48d167c", "address": "fa:16:3e:b0:3a:09", "network": {"id": "e16cdd31-8271-4c70-90fd-3943b04a95c9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-902402066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "092fc5912cd4449895f1c4d1318dd390", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6165b88-87", "ovs_interfaceid": "b6165b88-87f9-4c10-bc5c-fc58b48d167c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.534860] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:3a:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a55f45a-d631-4ebc-b73b-8a30bd0a32a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6165b88-87f9-4c10-bc5c-fc58b48d167c', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.543657] env[63355]: DEBUG oslo.service.loopingcall [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.549630] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.550994] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-481373dc-5b8e-4341-9897-c8bbd1ab4d9e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.581553] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349740, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.585794] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.585794] env[63355]: value = "task-1349743" [ 922.585794] env[63355]: _type = "Task" [ 922.585794] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.599394] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349743, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.779913] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815cd450-70f5-4c3f-994e-7ebf6cda3ecc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.791959] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0251b649-9e39-4597-8e39-f7389a57d6d0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.802143] env[63355]: DEBUG oslo_vmware.api [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349741, 'name': PowerOffVM_Task, 'duration_secs': 0.435079} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.802512] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.802767] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 922.838399] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-496c8484-ac18-45e2-8c76-098beba0c53e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.841417] env[63355]: DEBUG nova.network.neutron [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Successfully updated port: 0500cf25-c654-4748-a58a-8e51bd919768 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.843199] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7f523d-044a-4208-bf12-73d861e23b35 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.849323] env[63355]: DEBUG nova.compute.manager [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Received event network-changed-b6165b88-87f9-4c10-bc5c-fc58b48d167c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.849548] env[63355]: DEBUG nova.compute.manager [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Refreshing instance network info cache due to event network-changed-b6165b88-87f9-4c10-bc5c-fc58b48d167c. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 922.849788] env[63355]: DEBUG oslo_concurrency.lockutils [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] Acquiring lock "refresh_cache-199bc488-2e5d-4cea-aefb-ddd35ecc7a30" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.849949] env[63355]: DEBUG oslo_concurrency.lockutils [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] Acquired lock "refresh_cache-199bc488-2e5d-4cea-aefb-ddd35ecc7a30" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.850169] env[63355]: DEBUG nova.network.neutron [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Refreshing network info cache for port b6165b88-87f9-4c10-bc5c-fc58b48d167c {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 922.855664] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9de6c8a-0ab4-4f63-8b47-8a2e56280f61 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.876728] env[63355]: DEBUG nova.compute.provider_tree [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 922.882752] env[63355]: DEBUG nova.compute.manager [req-d2bc2679-642f-4803-9e90-0b5de87da16e req-f853e58a-fce8-4944-97bd-a325947b483e service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Received event network-vif-plugged-0500cf25-c654-4748-a58a-8e51bd919768 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.882752] env[63355]: DEBUG oslo_concurrency.lockutils [req-d2bc2679-642f-4803-9e90-0b5de87da16e req-f853e58a-fce8-4944-97bd-a325947b483e service nova] Acquiring lock "ecbacdb3-516e-43c3-96e3-4961b76565ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.882752] env[63355]: DEBUG oslo_concurrency.lockutils [req-d2bc2679-642f-4803-9e90-0b5de87da16e req-f853e58a-fce8-4944-97bd-a325947b483e service nova] Lock "ecbacdb3-516e-43c3-96e3-4961b76565ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.882752] env[63355]: DEBUG oslo_concurrency.lockutils [req-d2bc2679-642f-4803-9e90-0b5de87da16e req-f853e58a-fce8-4944-97bd-a325947b483e service nova] Lock "ecbacdb3-516e-43c3-96e3-4961b76565ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.882752] env[63355]: DEBUG nova.compute.manager [req-d2bc2679-642f-4803-9e90-0b5de87da16e req-f853e58a-fce8-4944-97bd-a325947b483e service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] No waiting events found dispatching network-vif-plugged-0500cf25-c654-4748-a58a-8e51bd919768 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 922.883082] env[63355]: WARNING nova.compute.manager [req-d2bc2679-642f-4803-9e90-0b5de87da16e req-f853e58a-fce8-4944-97bd-a325947b483e service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Received unexpected event network-vif-plugged-0500cf25-c654-4748-a58a-8e51bd919768 for instance with vm_state building and task_state block_device_mapping. [ 922.883082] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 922.883082] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 922.883082] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Deleting the datastore file [datastore1] b4b09b1d-680e-47b8-aa8a-9b3d9167824d {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.883613] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34b1393c-4c89-4248-a934-cd47fc383860 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.892030] env[63355]: DEBUG nova.compute.manager [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 922.892030] env[63355]: DEBUG oslo_vmware.api [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349735, 'name': RemoveSnapshot_Task, 'duration_secs': 1.132212} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.892637] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 922.892637] env[63355]: INFO nova.compute.manager [None req-efcaa8b7-7a4a-4275-a70a-2631a3f80923 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Took 18.25 seconds to snapshot the instance on the hypervisor. [ 922.902838] env[63355]: DEBUG oslo_vmware.api [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for the task: (returnval){ [ 922.902838] env[63355]: value = "task-1349745" [ 922.902838] env[63355]: _type = "Task" [ 922.902838] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.915094] env[63355]: DEBUG oslo_vmware.api [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349745, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.938685] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349742, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104154} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.938986] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.939818] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68db16b-1f75-4c69-9402-ebe24a07c6b8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.965928] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] dc56b350-cee7-49c4-9712-8c8c29146ff2/dc56b350-cee7-49c4-9712-8c8c29146ff2.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.966490] env[63355]: DEBUG nova.network.neutron [-] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.970519] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f27dce8-db78-4f4c-b7a3-fe5f98692fef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.985428] env[63355]: INFO nova.compute.manager [-] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Took 0.51 seconds to deallocate network for instance. [ 922.996444] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 922.996444] env[63355]: value = "task-1349746" [ 922.996444] env[63355]: _type = "Task" [ 922.996444] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.007495] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349746, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.080876] env[63355]: DEBUG oslo_vmware.api [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349740, 'name': PowerOnVM_Task, 'duration_secs': 0.585001} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.081180] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.081474] env[63355]: INFO nova.compute.manager [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Took 8.17 seconds to spawn the instance on the hypervisor. [ 923.081556] env[63355]: DEBUG nova.compute.manager [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.083080] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c8ff69-0427-4a0b-9423-c94e1ad5b66e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.104493] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349743, 'name': CreateVM_Task, 'duration_secs': 0.492153} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.104657] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.105373] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.105543] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.105954] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.106285] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38ffc947-9db2-4c4c-b050-421ff58b9fae {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.112737] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 923.112737] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5290cdf9-b248-7307-1d50-c535358c8735" [ 923.112737] env[63355]: _type = "Task" [ 923.112737] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.121334] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5290cdf9-b248-7307-1d50-c535358c8735, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.150803] env[63355]: DEBUG oslo_concurrency.lockutils [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "74071e73-10fa-4dcb-aa15-91303b2278fb" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.151115] env[63355]: DEBUG oslo_concurrency.lockutils [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "74071e73-10fa-4dcb-aa15-91303b2278fb" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.151353] env[63355]: INFO nova.compute.manager [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Rebooting instance [ 923.346934] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Acquiring lock "refresh_cache-ecbacdb3-516e-43c3-96e3-4961b76565ca" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.347300] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Acquired lock "refresh_cache-ecbacdb3-516e-43c3-96e3-4961b76565ca" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.347300] env[63355]: DEBUG nova.network.neutron [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.419481] env[63355]: DEBUG oslo_vmware.api [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Task: {'id': task-1349745, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176223} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.419481] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.419481] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 923.419481] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 923.419640] env[63355]: INFO nova.compute.manager [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 923.419943] env[63355]: DEBUG oslo.service.loopingcall [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.421026] env[63355]: DEBUG nova.scheduler.client.report [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Updated inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b with generation 91 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 923.421327] env[63355]: DEBUG nova.compute.provider_tree [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Updating resource provider 47abb610-db7e-4770-911d-187dd075ef8b generation from 91 to 92 during operation: update_inventory {{(pid=63355) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 923.421558] env[63355]: DEBUG nova.compute.provider_tree [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 923.426252] env[63355]: DEBUG nova.compute.manager [-] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.426415] env[63355]: DEBUG nova.network.neutron [-] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 923.430905] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.455011] env[63355]: DEBUG nova.network.neutron [-] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.492601] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.511453] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349746, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.608331] env[63355]: INFO nova.compute.manager [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Took 32.04 seconds to build instance. [ 923.624534] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5290cdf9-b248-7307-1d50-c535358c8735, 'name': SearchDatastore_Task, 'duration_secs': 0.013434} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.624889] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.625188] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.625468] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.625617] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.626608] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.626608] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e06f4214-3010-47cc-ab2b-116579ac94b9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.644514] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.644854] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.646745] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3560eb8-70d7-446b-9d9e-a71badaea63c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.661957] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 923.661957] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d59b64-3e61-b190-d4bf-a4f82f191a49" [ 923.661957] env[63355]: _type = "Task" [ 923.661957] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.677164] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d59b64-3e61-b190-d4bf-a4f82f191a49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.679051] env[63355]: DEBUG oslo_concurrency.lockutils [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.680068] env[63355]: DEBUG oslo_concurrency.lockutils [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquired lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.680068] env[63355]: DEBUG nova.network.neutron [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.738998] env[63355]: DEBUG nova.network.neutron [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Updated VIF entry in instance network info cache for port b6165b88-87f9-4c10-bc5c-fc58b48d167c. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 923.739404] env[63355]: DEBUG nova.network.neutron [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Updating instance_info_cache with network_info: [{"id": "b6165b88-87f9-4c10-bc5c-fc58b48d167c", "address": "fa:16:3e:b0:3a:09", "network": {"id": "e16cdd31-8271-4c70-90fd-3943b04a95c9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-902402066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "092fc5912cd4449895f1c4d1318dd390", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a55f45a-d631-4ebc-b73b-8a30bd0a32a8", "external-id": "nsx-vlan-transportzone-303", "segmentation_id": 303, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6165b88-87", "ovs_interfaceid": "b6165b88-87f9-4c10-bc5c-fc58b48d167c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.903720] env[63355]: DEBUG nova.network.neutron [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.932487] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.642s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.932997] env[63355]: DEBUG nova.compute.manager [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 923.935935] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.642s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.940304] env[63355]: INFO nova.compute.claims [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.959258] env[63355]: DEBUG nova.network.neutron [-] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.008923] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349746, 'name': ReconfigVM_Task, 'duration_secs': 0.551458} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.009261] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Reconfigured VM instance instance-00000047 to attach disk [datastore2] dc56b350-cee7-49c4-9712-8c8c29146ff2/dc56b350-cee7-49c4-9712-8c8c29146ff2.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.011144] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd4d99a0-b568-46df-a1bb-59c7c598d83e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.023554] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 924.023554] env[63355]: value = "task-1349747" [ 924.023554] env[63355]: _type = "Task" [ 924.023554] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.039196] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349747, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.101831] env[63355]: DEBUG nova.network.neutron [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Updating instance_info_cache with network_info: [{"id": "0500cf25-c654-4748-a58a-8e51bd919768", "address": "fa:16:3e:68:a5:c4", "network": {"id": "cca3f41d-524d-4176-871b-a2e826353fed", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1431855559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c787ea7464c54aa9a443b4945e8cb5df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0500cf25-c6", "ovs_interfaceid": "0500cf25-c654-4748-a58a-8e51bd919768", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.110329] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a4089c86-2295-416d-822d-acab16783551 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.082s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.147621] env[63355]: DEBUG nova.compute.manager [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 924.148244] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.148483] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.148652] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.148847] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.148997] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.149182] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.149400] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.149562] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.150365] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.150552] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.150735] env[63355]: DEBUG nova.virt.hardware [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.151640] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-758d0e1c-7372-4529-8a2d-2223bbf2b6bd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.161670] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f7cf62-3a35-447e-88b4-a070e8e31177 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.175384] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d59b64-3e61-b190-d4bf-a4f82f191a49, 'name': SearchDatastore_Task, 'duration_secs': 0.014011} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.186814] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0180e364-5e72-4a80-8d61-fbd7e64a4c01 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.194216] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 924.194216] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525b38c2-7a05-dfd9-f246-9f85167f6160" [ 924.194216] env[63355]: _type = "Task" [ 924.194216] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.204010] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525b38c2-7a05-dfd9-f246-9f85167f6160, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.241899] env[63355]: DEBUG oslo_concurrency.lockutils [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] Releasing lock "refresh_cache-199bc488-2e5d-4cea-aefb-ddd35ecc7a30" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.242143] env[63355]: DEBUG nova.compute.manager [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Received event network-changed-62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.242315] env[63355]: DEBUG nova.compute.manager [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Refreshing instance network info cache due to event network-changed-62623c8d-eb2e-4e47-87f9-52f1df6a7d7f. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 924.242499] env[63355]: DEBUG oslo_concurrency.lockutils [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] Acquiring lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.444181] env[63355]: DEBUG nova.network.neutron [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Updating instance_info_cache with network_info: [{"id": "62623c8d-eb2e-4e47-87f9-52f1df6a7d7f", "address": "fa:16:3e:9e:33:fc", "network": {"id": "47dda8ee-ad14-4eba-99a4-7a3afd76415c", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-564920032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87a296f0ecd94879ab2336899fc575bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "edd47158-6f4b-44a1-8e82-0411205ad299", "external-id": "nsx-vlan-transportzone-587", "segmentation_id": 587, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62623c8d-eb", "ovs_interfaceid": "62623c8d-eb2e-4e47-87f9-52f1df6a7d7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.449316] env[63355]: DEBUG nova.compute.utils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.451580] env[63355]: DEBUG nova.compute.manager [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 924.451580] env[63355]: DEBUG nova.network.neutron [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 924.460418] env[63355]: INFO nova.compute.manager [-] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Took 1.03 seconds to deallocate network for instance. [ 924.503018] env[63355]: DEBUG nova.policy [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfbc8f0af60940eeb0617698fb003a02', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd36c7cc647f741b38d717ba7c6458088', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 924.526440] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c2840fca-c129-48f9-8308-741b1bd1575c tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "interface-386e847e-967b-4247-9730-cdc5ac251474-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.526760] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c2840fca-c129-48f9-8308-741b1bd1575c tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-386e847e-967b-4247-9730-cdc5ac251474-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.527118] env[63355]: DEBUG nova.objects.instance [None req-c2840fca-c129-48f9-8308-741b1bd1575c tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'flavor' on Instance uuid 386e847e-967b-4247-9730-cdc5ac251474 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.538846] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349747, 'name': Rename_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.604829] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Releasing lock "refresh_cache-ecbacdb3-516e-43c3-96e3-4961b76565ca" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.605307] env[63355]: DEBUG nova.compute.manager [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Instance network_info: |[{"id": "0500cf25-c654-4748-a58a-8e51bd919768", "address": "fa:16:3e:68:a5:c4", "network": {"id": "cca3f41d-524d-4176-871b-a2e826353fed", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1431855559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c787ea7464c54aa9a443b4945e8cb5df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0500cf25-c6", "ovs_interfaceid": "0500cf25-c654-4748-a58a-8e51bd919768", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 924.606285] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:a5:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fdd0624-2edb-4733-8284-225815c07f73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0500cf25-c654-4748-a58a-8e51bd919768', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.618877] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Creating folder: Project (c787ea7464c54aa9a443b4945e8cb5df). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 924.619500] env[63355]: DEBUG nova.compute.manager [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 924.623381] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28b4b0fc-ccef-4c1d-9785-aa8841dfe38b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.640528] env[63355]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 924.640796] env[63355]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63355) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 924.641728] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Folder already exists: Project (c787ea7464c54aa9a443b4945e8cb5df). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 924.642141] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Creating folder: Instances. Parent ref: group-v287628. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 924.642794] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-babd4ef6-5040-4ea0-ade5-41ddce0fbd0c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.655568] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Created folder: Instances in parent group-v287628. [ 924.655845] env[63355]: DEBUG oslo.service.loopingcall [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.656053] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 924.656286] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-478435a0-a3e7-4371-b1bb-75a79d234c6d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.678804] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.678804] env[63355]: value = "task-1349750" [ 924.678804] env[63355]: _type = "Task" [ 924.678804] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.688328] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349750, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.704785] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525b38c2-7a05-dfd9-f246-9f85167f6160, 'name': SearchDatastore_Task, 'duration_secs': 0.018316} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.705145] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.705446] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 199bc488-2e5d-4cea-aefb-ddd35ecc7a30/199bc488-2e5d-4cea-aefb-ddd35ecc7a30.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 924.705759] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f2246400-c6f9-45f4-a540-4060bdf79f02 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.716515] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 924.716515] env[63355]: value = "task-1349751" [ 924.716515] env[63355]: _type = "Task" [ 924.716515] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.725436] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.955317] env[63355]: DEBUG oslo_concurrency.lockutils [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Releasing lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.957571] env[63355]: DEBUG nova.compute.manager [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 924.960122] env[63355]: DEBUG oslo_concurrency.lockutils [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] Acquired lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.960122] env[63355]: DEBUG nova.network.neutron [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Refreshing network info cache for port 62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 924.961166] env[63355]: DEBUG nova.compute.manager [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.962290] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749f742a-73a6-409a-86c0-2d83157e98e0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.970994] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.038245] env[63355]: DEBUG nova.objects.instance [None req-c2840fca-c129-48f9-8308-741b1bd1575c tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'pci_requests' on Instance uuid 386e847e-967b-4247-9730-cdc5ac251474 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.046332] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349747, 'name': Rename_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.149568] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.194250] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349750, 'name': CreateVM_Task, 'duration_secs': 0.440609} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.197327] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 925.198501] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'mount_device': '/dev/sda', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287631', 'volume_id': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'name': 'volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ecbacdb3-516e-43c3-96e3-4961b76565ca', 'attached_at': '', 'detached_at': '', 'volume_id': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'serial': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1'}, 'attachment_id': 'd2995d0e-77cd-4387-be67-509b1ccf3a99', 'guest_format': None, 'delete_on_termination': True, 'disk_bus': None, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=63355) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 925.198734] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Root volume attach. Driver type: vmdk {{(pid=63355) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 925.199608] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2f8404-b8eb-43c9-a822-e51f1eca7066 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.211981] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982427da-6407-4290-9e86-269b2713a294 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.224711] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e6e136-461b-4ce2-bc23-c4e333b96522 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.232575] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349751, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.239733] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-6ba28323-d4c2-44c6-9e1f-7db23110384a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.250413] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for the task: (returnval){ [ 925.250413] env[63355]: value = "task-1349752" [ 925.250413] env[63355]: _type = "Task" [ 925.250413] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.262730] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349752, 'name': RelocateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.367635] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5dd1f6f-bc2e-4072-bb0a-42f76f91a752 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.377373] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680ed971-ffc4-44ec-a948-656bcad2f207 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.411477] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467b6ce5-b616-456f-898a-a6f1c712f662 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.420736] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5df753f-f591-425a-ad0a-659dcb93e8ce {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.441496] env[63355]: DEBUG nova.compute.provider_tree [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.478259] env[63355]: DEBUG nova.network.neutron [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Successfully created port: 8c82e15c-a3e6-4ec2-b00e-cc57c343b635 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.544681] env[63355]: DEBUG nova.objects.base [None req-c2840fca-c129-48f9-8308-741b1bd1575c tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Object Instance<386e847e-967b-4247-9730-cdc5ac251474> lazy-loaded attributes: flavor,pci_requests {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 925.544928] env[63355]: DEBUG nova.network.neutron [None req-c2840fca-c129-48f9-8308-741b1bd1575c tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 925.546953] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349747, 'name': Rename_Task, 'duration_secs': 1.242789} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.549615] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.550424] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3541f6f4-e1b7-49af-873b-4d044550ae6d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.563270] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 925.563270] env[63355]: value = "task-1349753" [ 925.563270] env[63355]: _type = "Task" [ 925.563270] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.576479] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.677911] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c2840fca-c129-48f9-8308-741b1bd1575c tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-386e847e-967b-4247-9730-cdc5ac251474-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.151s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.732813] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349751, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.686632} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.733112] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 199bc488-2e5d-4cea-aefb-ddd35ecc7a30/199bc488-2e5d-4cea-aefb-ddd35ecc7a30.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.733341] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.733627] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d35d40c6-9f59-4ab5-a6c5-200ceed912b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.744324] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 925.744324] env[63355]: value = "task-1349754" [ 925.744324] env[63355]: _type = "Task" [ 925.744324] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.758366] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349754, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.767917] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349752, 'name': RelocateVM_Task} progress is 40%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.769871] env[63355]: DEBUG nova.network.neutron [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Updated VIF entry in instance network info cache for port 62623c8d-eb2e-4e47-87f9-52f1df6a7d7f. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 925.770257] env[63355]: DEBUG nova.network.neutron [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Updating instance_info_cache with network_info: [{"id": "62623c8d-eb2e-4e47-87f9-52f1df6a7d7f", "address": "fa:16:3e:9e:33:fc", "network": {"id": "47dda8ee-ad14-4eba-99a4-7a3afd76415c", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-564920032-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87a296f0ecd94879ab2336899fc575bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "edd47158-6f4b-44a1-8e82-0411205ad299", "external-id": "nsx-vlan-transportzone-587", "segmentation_id": 587, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62623c8d-eb", "ovs_interfaceid": "62623c8d-eb2e-4e47-87f9-52f1df6a7d7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.949385] env[63355]: DEBUG nova.scheduler.client.report [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.972022] env[63355]: DEBUG nova.compute.manager [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 925.985882] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3593719c-1be8-4544-b9d3-8597011809d1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.000983] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Doing hard reboot of VM {{(pid=63355) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 926.007034] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-43e58026-c9a1-473f-9b78-ef125d9b7aa5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.020334] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.021590] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.024022] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.024022] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.024022] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.024022] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.024022] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.024371] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.024371] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.024371] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.024749] env[63355]: DEBUG nova.virt.hardware [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.026211] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e6e9a9-c752-4411-b8bb-d18ad20668c8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.032622] env[63355]: DEBUG oslo_vmware.api [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 926.032622] env[63355]: value = "task-1349755" [ 926.032622] env[63355]: _type = "Task" [ 926.032622] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.043013] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f9f2e3-d012-4cda-9141-436d5fe56b54 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.051429] env[63355]: DEBUG oslo_vmware.api [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349755, 'name': ResetVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.083365] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.084921] env[63355]: DEBUG nova.compute.manager [req-339a9950-15ff-40fb-b403-f1a9faf2c78e req-a1918ee2-bd63-4c22-9069-f8ab3a89b7dc service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Received event network-changed-0500cf25-c654-4748-a58a-8e51bd919768 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.087395] env[63355]: DEBUG nova.compute.manager [req-339a9950-15ff-40fb-b403-f1a9faf2c78e req-a1918ee2-bd63-4c22-9069-f8ab3a89b7dc service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Refreshing instance network info cache due to event network-changed-0500cf25-c654-4748-a58a-8e51bd919768. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 926.087395] env[63355]: DEBUG oslo_concurrency.lockutils [req-339a9950-15ff-40fb-b403-f1a9faf2c78e req-a1918ee2-bd63-4c22-9069-f8ab3a89b7dc service nova] Acquiring lock "refresh_cache-ecbacdb3-516e-43c3-96e3-4961b76565ca" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.087395] env[63355]: DEBUG oslo_concurrency.lockutils [req-339a9950-15ff-40fb-b403-f1a9faf2c78e req-a1918ee2-bd63-4c22-9069-f8ab3a89b7dc service nova] Acquired lock "refresh_cache-ecbacdb3-516e-43c3-96e3-4961b76565ca" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.087395] env[63355]: DEBUG nova.network.neutron [req-339a9950-15ff-40fb-b403-f1a9faf2c78e req-a1918ee2-bd63-4c22-9069-f8ab3a89b7dc service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Refreshing network info cache for port 0500cf25-c654-4748-a58a-8e51bd919768 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 926.263120] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349752, 'name': RelocateVM_Task} progress is 51%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.266064] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349754, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.292997} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.266388] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.267305] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff29ecb-c4fe-4f13-8267-a88955da34e8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.273242] env[63355]: DEBUG oslo_concurrency.lockutils [req-b6da6d22-42aa-4b95-9a4d-b697de896122 req-c29c08b2-f361-4c79-9fb7-3328cd96ede9 service nova] Releasing lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.295458] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 199bc488-2e5d-4cea-aefb-ddd35ecc7a30/199bc488-2e5d-4cea-aefb-ddd35ecc7a30.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.296962] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ec3ed0e-4aa0-4c67-b56c-cc2ccd48e7af {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.312340] env[63355]: DEBUG nova.compute.manager [req-0c5f42c5-39a8-4ba7-9ba2-62cd934fb92c req-2956a3f5-5528-4bed-92f9-40de153b077f service nova] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Received event network-vif-deleted-00f30816-4a44-44bc-9eb3-ca224f3da164 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.323829] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 926.323829] env[63355]: value = "task-1349756" [ 926.323829] env[63355]: _type = "Task" [ 926.323829] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.335570] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349756, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.459041] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.459339] env[63355]: DEBUG nova.compute.manager [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 926.462019] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.257s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.464236] env[63355]: INFO nova.compute.claims [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.547432] env[63355]: DEBUG oslo_vmware.api [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349755, 'name': ResetVM_Task, 'duration_secs': 0.099896} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.547778] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Did hard reboot of VM {{(pid=63355) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 926.547950] env[63355]: DEBUG nova.compute.manager [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.548806] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5487b1b-4408-41d1-8196-2c0a8dcd5a3a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.582746] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.770303] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349752, 'name': RelocateVM_Task} progress is 63%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.839477] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349756, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.012115] env[63355]: DEBUG nova.compute.utils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 927.017957] env[63355]: DEBUG nova.compute.manager [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 927.018229] env[63355]: DEBUG nova.network.neutron [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 927.067148] env[63355]: DEBUG oslo_concurrency.lockutils [None req-94b17694-d89f-422d-8990-cd38828c8ac1 tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "74071e73-10fa-4dcb-aa15-91303b2278fb" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.915s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.084289] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349753, 'name': PowerOnVM_Task} progress is 71%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.110276] env[63355]: DEBUG nova.policy [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc8822a59d6b41e19da70d634da7b67b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c6bd8f13718c44e7bc41c7974ac90d28', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 927.227901] env[63355]: DEBUG nova.network.neutron [req-339a9950-15ff-40fb-b403-f1a9faf2c78e req-a1918ee2-bd63-4c22-9069-f8ab3a89b7dc service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Updated VIF entry in instance network info cache for port 0500cf25-c654-4748-a58a-8e51bd919768. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 927.228393] env[63355]: DEBUG nova.network.neutron [req-339a9950-15ff-40fb-b403-f1a9faf2c78e req-a1918ee2-bd63-4c22-9069-f8ab3a89b7dc service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Updating instance_info_cache with network_info: [{"id": "0500cf25-c654-4748-a58a-8e51bd919768", "address": "fa:16:3e:68:a5:c4", "network": {"id": "cca3f41d-524d-4176-871b-a2e826353fed", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1431855559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c787ea7464c54aa9a443b4945e8cb5df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0500cf25-c6", "ovs_interfaceid": "0500cf25-c654-4748-a58a-8e51bd919768", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.270319] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349752, 'name': RelocateVM_Task} progress is 76%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.344020] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349756, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.488816] env[63355]: DEBUG nova.network.neutron [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Successfully created port: de9150b0-5d81-4397-a0d7-58c9a471d791 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 927.521035] env[63355]: DEBUG nova.compute.manager [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 927.588962] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349753, 'name': PowerOnVM_Task} progress is 71%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.688578] env[63355]: DEBUG nova.compute.manager [req-10628425-afbf-475f-b229-c1ee1eebb6ab req-a13ca012-7ad2-4c22-99b0-6170723552ab service nova] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Received event network-vif-plugged-8c82e15c-a3e6-4ec2-b00e-cc57c343b635 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.688578] env[63355]: DEBUG oslo_concurrency.lockutils [req-10628425-afbf-475f-b229-c1ee1eebb6ab req-a13ca012-7ad2-4c22-99b0-6170723552ab service nova] Acquiring lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.688825] env[63355]: DEBUG oslo_concurrency.lockutils [req-10628425-afbf-475f-b229-c1ee1eebb6ab req-a13ca012-7ad2-4c22-99b0-6170723552ab service nova] Lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.689008] env[63355]: DEBUG oslo_concurrency.lockutils [req-10628425-afbf-475f-b229-c1ee1eebb6ab req-a13ca012-7ad2-4c22-99b0-6170723552ab service nova] Lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.689197] env[63355]: DEBUG nova.compute.manager [req-10628425-afbf-475f-b229-c1ee1eebb6ab req-a13ca012-7ad2-4c22-99b0-6170723552ab service nova] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] No waiting events found dispatching network-vif-plugged-8c82e15c-a3e6-4ec2-b00e-cc57c343b635 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 927.689369] env[63355]: WARNING nova.compute.manager [req-10628425-afbf-475f-b229-c1ee1eebb6ab req-a13ca012-7ad2-4c22-99b0-6170723552ab service nova] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Received unexpected event network-vif-plugged-8c82e15c-a3e6-4ec2-b00e-cc57c343b635 for instance with vm_state building and task_state spawning. [ 927.733770] env[63355]: DEBUG oslo_concurrency.lockutils [req-339a9950-15ff-40fb-b403-f1a9faf2c78e req-a1918ee2-bd63-4c22-9069-f8ab3a89b7dc service nova] Releasing lock "refresh_cache-ecbacdb3-516e-43c3-96e3-4961b76565ca" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.772848] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349752, 'name': RelocateVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.836302] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "interface-386e847e-967b-4247-9730-cdc5ac251474-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.836599] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-386e847e-967b-4247-9730-cdc5ac251474-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.836986] env[63355]: DEBUG nova.objects.instance [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'flavor' on Instance uuid 386e847e-967b-4247-9730-cdc5ac251474 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.842150] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349756, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.869325] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4883c5f-7600-42dd-a183-b3f6dba88d52 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.879991] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b4aa02-7cb2-49f2-a0e3-f57ed6c9f307 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.920057] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46310334-1789-4126-98bc-0806f2fbc7ce {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.929928] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2d6f22-8371-4611-b4c5-3086119f79bc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.946897] env[63355]: DEBUG nova.compute.provider_tree [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.963333] env[63355]: DEBUG nova.network.neutron [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Successfully updated port: 8c82e15c-a3e6-4ec2-b00e-cc57c343b635 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.088685] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349753, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.266143] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349752, 'name': RelocateVM_Task} progress is 97%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.297744] env[63355]: DEBUG oslo_vmware.rw_handles [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c3cc-d8d0-75ae-83bf-d0e0b7dc574e/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 928.298792] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4828203f-a66b-409e-9693-3941ff7c6ca4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.305828] env[63355]: DEBUG oslo_vmware.rw_handles [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c3cc-d8d0-75ae-83bf-d0e0b7dc574e/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 928.306035] env[63355]: ERROR oslo_vmware.rw_handles [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c3cc-d8d0-75ae-83bf-d0e0b7dc574e/disk-0.vmdk due to incomplete transfer. [ 928.306284] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0cc49637-b47e-4d20-9805-6038f51e7eb4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.314187] env[63355]: DEBUG oslo_vmware.rw_handles [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c3cc-d8d0-75ae-83bf-d0e0b7dc574e/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 928.314406] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Uploaded image 98bf2e17-09e0-42e8-ab0d-b84472dc2ae4 to the Glance image server {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 928.316179] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 928.316437] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-89142ab9-9225-44c8-80ab-61572076c9bb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.324265] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 928.324265] env[63355]: value = "task-1349757" [ 928.324265] env[63355]: _type = "Task" [ 928.324265] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.334657] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349757, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.342850] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349756, 'name': ReconfigVM_Task, 'duration_secs': 1.750311} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.343139] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 199bc488-2e5d-4cea-aefb-ddd35ecc7a30/199bc488-2e5d-4cea-aefb-ddd35ecc7a30.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.344065] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1f6fe36-d5b2-4488-b936-55a5fe0ccd7d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.352394] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 928.352394] env[63355]: value = "task-1349758" [ 928.352394] env[63355]: _type = "Task" [ 928.352394] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.363933] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349758, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.453688] env[63355]: DEBUG nova.scheduler.client.report [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.465685] env[63355]: DEBUG nova.objects.instance [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'pci_requests' on Instance uuid 386e847e-967b-4247-9730-cdc5ac251474 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.466943] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "refresh_cache-d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.467394] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquired lock "refresh_cache-d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.467394] env[63355]: DEBUG nova.network.neutron [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 928.533776] env[63355]: DEBUG nova.compute.manager [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 928.555114] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.555365] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.555524] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.555749] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.555927] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.556098] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.556309] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.556470] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.556635] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.556830] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.557027] env[63355]: DEBUG nova.virt.hardware [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.557879] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c9a0fd-c300-4146-814e-eab0e17f6371 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.566280] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d0c284-74fe-4790-a2ac-62183041892f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.590006] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349753, 'name': PowerOnVM_Task, 'duration_secs': 2.623681} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.590284] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.590484] env[63355]: INFO nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Took 11.20 seconds to spawn the instance on the hypervisor. [ 928.590661] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.591404] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c68b08b-1d01-4017-8ae9-6dc56b939923 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.768182] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349752, 'name': RelocateVM_Task} progress is 97%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.835450] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349757, 'name': Destroy_Task, 'duration_secs': 0.383604} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.836218] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Destroyed the VM [ 928.836218] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 928.836434] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c7b8777d-c65c-4aa7-bf4f-1ca9f3613a49 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.844654] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 928.844654] env[63355]: value = "task-1349759" [ 928.844654] env[63355]: _type = "Task" [ 928.844654] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.854363] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349759, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.862384] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349758, 'name': Rename_Task, 'duration_secs': 0.170608} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.862693] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.862944] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7f2104b-ad80-4019-9bd7-133680dcc80b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.869218] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 928.869218] env[63355]: value = "task-1349760" [ 928.869218] env[63355]: _type = "Task" [ 928.869218] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.877640] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349760, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.959354] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.960030] env[63355]: DEBUG nova.compute.manager [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 928.963627] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.306s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.964036] env[63355]: DEBUG nova.objects.instance [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lazy-loading 'resources' on Instance uuid b6056441-9ee4-484f-a1d2-077546f2c581 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.969414] env[63355]: DEBUG nova.objects.base [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Object Instance<386e847e-967b-4247-9730-cdc5ac251474> lazy-loaded attributes: flavor,pci_requests {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 928.969611] env[63355]: DEBUG nova.network.neutron [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 929.022709] env[63355]: DEBUG nova.network.neutron [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 929.036018] env[63355]: DEBUG nova.policy [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77243643ea724b72858a8682a2a054a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d47b671ea9c429391cbdae7e24adadf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 929.041341] env[63355]: DEBUG nova.compute.manager [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Received event network-changed-976ab0e0-5cea-4265-bc51-9b5cd29ed954 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.041530] env[63355]: DEBUG nova.compute.manager [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Refreshing instance network info cache due to event network-changed-976ab0e0-5cea-4265-bc51-9b5cd29ed954. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.042302] env[63355]: DEBUG oslo_concurrency.lockutils [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] Acquiring lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.042302] env[63355]: DEBUG oslo_concurrency.lockutils [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] Acquired lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.042302] env[63355]: DEBUG nova.network.neutron [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Refreshing network info cache for port 976ab0e0-5cea-4265-bc51-9b5cd29ed954 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.109458] env[63355]: DEBUG nova.network.neutron [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Successfully updated port: de9150b0-5d81-4397-a0d7-58c9a471d791 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 929.116127] env[63355]: INFO nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Took 35.71 seconds to build instance. [ 929.215010] env[63355]: DEBUG nova.network.neutron [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Updating instance_info_cache with network_info: [{"id": "8c82e15c-a3e6-4ec2-b00e-cc57c343b635", "address": "fa:16:3e:10:ad:3a", "network": {"id": "9b1d268a-c013-4c6c-a69b-326a9876ad52", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-137926478-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d36c7cc647f741b38d717ba7c6458088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f67a2790-f2b0-4d03-b606-0bfaee7a4229", "external-id": "nsx-vlan-transportzone-187", "segmentation_id": 187, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c82e15c-a3", "ovs_interfaceid": "8c82e15c-a3e6-4ec2-b00e-cc57c343b635", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.269339] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349752, 'name': RelocateVM_Task} progress is 98%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.355017] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349759, 'name': RemoveSnapshot_Task} progress is 43%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.382282] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349760, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.440116] env[63355]: DEBUG nova.network.neutron [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Successfully created port: 678307c8-eb2a-409d-a339-e40d5fd2ee50 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.469920] env[63355]: DEBUG nova.compute.utils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 929.474630] env[63355]: DEBUG nova.compute.manager [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 929.474886] env[63355]: DEBUG nova.network.neutron [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 929.523372] env[63355]: DEBUG nova.policy [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fdbd89aac89e4d58bbfd5b549d42a27f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '744ce7dc0ab4463fa461cbac87e51904', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 929.615951] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquiring lock "refresh_cache-a408e19a-0960-430d-8550-0a304c63da61" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.616208] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquired lock "refresh_cache-a408e19a-0960-430d-8550-0a304c63da61" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.616366] env[63355]: DEBUG nova.network.neutron [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 929.618106] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "dc56b350-cee7-49c4-9712-8c8c29146ff2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.400s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.720965] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Releasing lock "refresh_cache-d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.723763] env[63355]: DEBUG nova.compute.manager [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Instance network_info: |[{"id": "8c82e15c-a3e6-4ec2-b00e-cc57c343b635", "address": "fa:16:3e:10:ad:3a", "network": {"id": "9b1d268a-c013-4c6c-a69b-326a9876ad52", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-137926478-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d36c7cc647f741b38d717ba7c6458088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f67a2790-f2b0-4d03-b606-0bfaee7a4229", "external-id": "nsx-vlan-transportzone-187", "segmentation_id": 187, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c82e15c-a3", "ovs_interfaceid": "8c82e15c-a3e6-4ec2-b00e-cc57c343b635", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 929.725037] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:ad:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f67a2790-f2b0-4d03-b606-0bfaee7a4229', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c82e15c-a3e6-4ec2-b00e-cc57c343b635', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 929.733751] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Creating folder: Project (d36c7cc647f741b38d717ba7c6458088). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 929.737010] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ebd05cb2-b474-495d-88f2-928c8bb8918c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.756183] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Created folder: Project (d36c7cc647f741b38d717ba7c6458088) in parent group-v287607. [ 929.756183] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Creating folder: Instances. Parent ref: group-v287687. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 929.756183] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b533ee5-f944-46a8-9932-c16dea856955 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.767719] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349752, 'name': RelocateVM_Task, 'duration_secs': 4.136788} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.771610] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 929.771842] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287631', 'volume_id': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'name': 'volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ecbacdb3-516e-43c3-96e3-4961b76565ca', 'attached_at': '', 'detached_at': '', 'volume_id': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'serial': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 929.772198] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Created folder: Instances in parent group-v287687. [ 929.772418] env[63355]: DEBUG oslo.service.loopingcall [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.773411] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eda1009-5c67-45bc-a867-f4e5e57f5466 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.775834] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 929.776388] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9c6f7e3-e54a-437e-8085-fe6119b77b42 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.813731] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c3f3b7-78cf-42d1-973c-98411e2a4c25 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.816475] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 929.816475] env[63355]: value = "task-1349763" [ 929.816475] env[63355]: _type = "Task" [ 929.816475] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.837251] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1/volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 929.844204] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a590e5ea-ee93-4457-9727-5fad7357a933 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.860578] env[63355]: DEBUG nova.compute.manager [req-9e512a43-1d26-4f1a-812e-27a8d5dccb4b req-eafda80b-98a4-4a19-b9d6-5c569224c546 service nova] [instance: a408e19a-0960-430d-8550-0a304c63da61] Received event network-vif-plugged-de9150b0-5d81-4397-a0d7-58c9a471d791 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.860817] env[63355]: DEBUG oslo_concurrency.lockutils [req-9e512a43-1d26-4f1a-812e-27a8d5dccb4b req-eafda80b-98a4-4a19-b9d6-5c569224c546 service nova] Acquiring lock "a408e19a-0960-430d-8550-0a304c63da61-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.861045] env[63355]: DEBUG oslo_concurrency.lockutils [req-9e512a43-1d26-4f1a-812e-27a8d5dccb4b req-eafda80b-98a4-4a19-b9d6-5c569224c546 service nova] Lock "a408e19a-0960-430d-8550-0a304c63da61-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.861219] env[63355]: DEBUG oslo_concurrency.lockutils [req-9e512a43-1d26-4f1a-812e-27a8d5dccb4b req-eafda80b-98a4-4a19-b9d6-5c569224c546 service nova] Lock "a408e19a-0960-430d-8550-0a304c63da61-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.861384] env[63355]: DEBUG nova.compute.manager [req-9e512a43-1d26-4f1a-812e-27a8d5dccb4b req-eafda80b-98a4-4a19-b9d6-5c569224c546 service nova] [instance: a408e19a-0960-430d-8550-0a304c63da61] No waiting events found dispatching network-vif-plugged-de9150b0-5d81-4397-a0d7-58c9a471d791 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.861544] env[63355]: WARNING nova.compute.manager [req-9e512a43-1d26-4f1a-812e-27a8d5dccb4b req-eafda80b-98a4-4a19-b9d6-5c569224c546 service nova] [instance: a408e19a-0960-430d-8550-0a304c63da61] Received unexpected event network-vif-plugged-de9150b0-5d81-4397-a0d7-58c9a471d791 for instance with vm_state building and task_state spawning. [ 929.866067] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349763, 'name': CreateVM_Task} progress is 15%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.879676] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349759, 'name': RemoveSnapshot_Task} progress is 98%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.880043] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for the task: (returnval){ [ 929.880043] env[63355]: value = "task-1349764" [ 929.880043] env[63355]: _type = "Task" [ 929.880043] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.882633] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2aacb90-2b75-4282-91a2-70886f4ea206 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.894127] env[63355]: DEBUG oslo_vmware.api [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349760, 'name': PowerOnVM_Task, 'duration_secs': 0.614859} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.895072] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.895348] env[63355]: INFO nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Took 10.03 seconds to spawn the instance on the hypervisor. [ 929.895544] env[63355]: DEBUG nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.896897] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b92714-7f78-4dae-bfa8-a8e14024ae84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.906126] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349764, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.908043] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ed19d8-5173-4425-84aa-6383f444c13a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.949248] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbf72dc-d045-4512-9b32-d01a5913ff80 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.952890] env[63355]: DEBUG nova.network.neutron [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Successfully created port: 43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.963628] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-706fe684-56bb-4352-82da-d903f4785219 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.979628] env[63355]: DEBUG nova.compute.manager [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 929.983246] env[63355]: DEBUG nova.compute.provider_tree [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 930.028013] env[63355]: DEBUG nova.network.neutron [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updated VIF entry in instance network info cache for port 976ab0e0-5cea-4265-bc51-9b5cd29ed954. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 930.028013] env[63355]: DEBUG nova.network.neutron [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance_info_cache with network_info: [{"id": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "address": "fa:16:3e:b3:26:57", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap976ab0e0-5c", "ovs_interfaceid": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.124885] env[63355]: DEBUG nova.compute.manager [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 930.172461] env[63355]: DEBUG nova.network.neutron [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 930.326989] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349763, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.373069] env[63355]: DEBUG nova.network.neutron [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Updating instance_info_cache with network_info: [{"id": "de9150b0-5d81-4397-a0d7-58c9a471d791", "address": "fa:16:3e:7d:8d:90", "network": {"id": "f0995df2-4470-43ed-8168-2d1324820fe0", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2109446375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6bd8f13718c44e7bc41c7974ac90d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde9150b0-5d", "ovs_interfaceid": "de9150b0-5d81-4397-a0d7-58c9a471d791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.378459] env[63355]: DEBUG oslo_vmware.api [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349759, 'name': RemoveSnapshot_Task, 'duration_secs': 1.30317} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.378839] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 930.379105] env[63355]: INFO nova.compute.manager [None req-e603ca4d-14b4-42d0-a143-2c826859454f tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Took 18.17 seconds to snapshot the instance on the hypervisor. [ 930.390988] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349764, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.426144] env[63355]: INFO nova.compute.manager [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Took 36.98 seconds to build instance. [ 930.534172] env[63355]: DEBUG oslo_concurrency.lockutils [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] Releasing lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.534445] env[63355]: DEBUG nova.compute.manager [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Received event network-changed-8c82e15c-a3e6-4ec2-b00e-cc57c343b635 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.534726] env[63355]: DEBUG nova.compute.manager [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Refreshing instance network info cache due to event network-changed-8c82e15c-a3e6-4ec2-b00e-cc57c343b635. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 930.534932] env[63355]: DEBUG oslo_concurrency.lockutils [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] Acquiring lock "refresh_cache-d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.534998] env[63355]: DEBUG oslo_concurrency.lockutils [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] Acquired lock "refresh_cache-d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.535130] env[63355]: DEBUG nova.network.neutron [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Refreshing network info cache for port 8c82e15c-a3e6-4ec2-b00e-cc57c343b635 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 930.537243] env[63355]: DEBUG nova.scheduler.client.report [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Updated inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b with generation 92 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 930.537488] env[63355]: DEBUG nova.compute.provider_tree [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Updating resource provider 47abb610-db7e-4770-911d-187dd075ef8b generation from 92 to 93 during operation: update_inventory {{(pid=63355) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 930.537673] env[63355]: DEBUG nova.compute.provider_tree [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 930.649273] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.774781] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "74071e73-10fa-4dcb-aa15-91303b2278fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.775367] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "74071e73-10fa-4dcb-aa15-91303b2278fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.775367] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "74071e73-10fa-4dcb-aa15-91303b2278fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.775367] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "74071e73-10fa-4dcb-aa15-91303b2278fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.775537] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "74071e73-10fa-4dcb-aa15-91303b2278fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.777849] env[63355]: INFO nova.compute.manager [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Terminating instance [ 930.779671] env[63355]: DEBUG nova.compute.manager [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.779891] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 930.780743] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a512378-f584-4ea2-9805-b148ddc7b77b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.790579] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.790835] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26b7393f-49a8-4237-8efe-06b417c5b87b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.798591] env[63355]: DEBUG oslo_vmware.api [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 930.798591] env[63355]: value = "task-1349765" [ 930.798591] env[63355]: _type = "Task" [ 930.798591] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.808329] env[63355]: DEBUG oslo_vmware.api [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349765, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.828088] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349763, 'name': CreateVM_Task, 'duration_secs': 0.6032} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.828291] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 930.828994] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.829198] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.829533] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.829805] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fe696e4-c889-4f6d-995d-ca7c1d88b9be {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.837020] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for the task: (returnval){ [ 930.837020] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52405a26-30b8-082e-63da-8bbe125d263a" [ 930.837020] env[63355]: _type = "Task" [ 930.837020] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.844772] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52405a26-30b8-082e-63da-8bbe125d263a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.880331] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Releasing lock "refresh_cache-a408e19a-0960-430d-8550-0a304c63da61" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.880331] env[63355]: DEBUG nova.compute.manager [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Instance network_info: |[{"id": "de9150b0-5d81-4397-a0d7-58c9a471d791", "address": "fa:16:3e:7d:8d:90", "network": {"id": "f0995df2-4470-43ed-8168-2d1324820fe0", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2109446375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6bd8f13718c44e7bc41c7974ac90d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde9150b0-5d", "ovs_interfaceid": "de9150b0-5d81-4397-a0d7-58c9a471d791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.880642] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:8d:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae18b41f-e73c-44f1-83dd-467c080944f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'de9150b0-5d81-4397-a0d7-58c9a471d791', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.888673] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Creating folder: Project (c6bd8f13718c44e7bc41c7974ac90d28). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 930.891393] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d8e596a6-e553-457f-9afb-68601ffd3155 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.904087] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349764, 'name': ReconfigVM_Task, 'duration_secs': 0.690758} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.904500] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Reconfigured VM instance instance-00000049 to attach disk [datastore1] volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1/volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.912071] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4564f9e9-fe2b-4c7f-90b1-8dcfaaa0c016 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.923862] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Created folder: Project (c6bd8f13718c44e7bc41c7974ac90d28) in parent group-v287607. [ 930.923862] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Creating folder: Instances. Parent ref: group-v287690. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 930.923862] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f920371c-b5c8-4a67-99be-4b9e807568a9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.928342] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4de72279-9e01-426f-8dc7-7ccf0021fb34 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.677s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.931651] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for the task: (returnval){ [ 930.931651] env[63355]: value = "task-1349768" [ 930.931651] env[63355]: _type = "Task" [ 930.931651] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.938875] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Created folder: Instances in parent group-v287690. [ 930.939119] env[63355]: DEBUG oslo.service.loopingcall [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.939899] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a408e19a-0960-430d-8550-0a304c63da61] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 930.940079] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-313fa270-d48d-487e-a498-07b9d0ccedf4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.958846] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349768, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.965331] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.965331] env[63355]: value = "task-1349769" [ 930.965331] env[63355]: _type = "Task" [ 930.965331] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.974883] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349769, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.994046] env[63355]: DEBUG nova.compute.manager [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 931.018382] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 931.018680] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 931.018820] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.019051] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.019269] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.019408] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.019660] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.019837] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.019929] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.020130] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.020364] env[63355]: DEBUG nova.virt.hardware [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.021297] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20aa8ba-b59b-46b1-857d-4e20515a7b9c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.034722] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e3d5e0-4f90-417e-a4d3-59a4afd10c9f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.046540] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.083s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.058684] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.073s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.061071] env[63355]: INFO nova.compute.claims [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.068323] env[63355]: INFO nova.scheduler.client.report [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Deleted allocations for instance b6056441-9ee4-484f-a1d2-077546f2c581 [ 931.307967] env[63355]: DEBUG nova.network.neutron [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Updated VIF entry in instance network info cache for port 8c82e15c-a3e6-4ec2-b00e-cc57c343b635. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 931.308388] env[63355]: DEBUG nova.network.neutron [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Updating instance_info_cache with network_info: [{"id": "8c82e15c-a3e6-4ec2-b00e-cc57c343b635", "address": "fa:16:3e:10:ad:3a", "network": {"id": "9b1d268a-c013-4c6c-a69b-326a9876ad52", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-137926478-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d36c7cc647f741b38d717ba7c6458088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f67a2790-f2b0-4d03-b606-0bfaee7a4229", "external-id": "nsx-vlan-transportzone-187", "segmentation_id": 187, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c82e15c-a3", "ovs_interfaceid": "8c82e15c-a3e6-4ec2-b00e-cc57c343b635", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.316454] env[63355]: DEBUG oslo_vmware.api [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349765, 'name': PowerOffVM_Task, 'duration_secs': 0.193278} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.316659] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 931.316852] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 931.320101] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e0df0f1-ac14-4584-8a6d-e0f72f5b4858 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.346316] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52405a26-30b8-082e-63da-8bbe125d263a, 'name': SearchDatastore_Task, 'duration_secs': 0.016689} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.346637] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.346893] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.347254] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.347370] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.347512] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.347785] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-506df108-22af-47ae-8a36-0f22426fe164 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.358549] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.358819] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 931.359563] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66f26bba-30c7-480a-b314-7e50ea00658e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.362699] env[63355]: DEBUG nova.network.neutron [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Successfully updated port: 678307c8-eb2a-409d-a339-e40d5fd2ee50 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 931.367782] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for the task: (returnval){ [ 931.367782] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5206922a-e4c2-b1d0-0879-34db2f3afd45" [ 931.367782] env[63355]: _type = "Task" [ 931.367782] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.378773] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5206922a-e4c2-b1d0-0879-34db2f3afd45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.398742] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 931.398742] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 931.398742] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Deleting the datastore file [datastore1] 74071e73-10fa-4dcb-aa15-91303b2278fb {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.398742] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6fa17274-ae05-44f3-9d2e-560c6c5ebd99 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.406825] env[63355]: DEBUG oslo_vmware.api [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for the task: (returnval){ [ 931.406825] env[63355]: value = "task-1349771" [ 931.406825] env[63355]: _type = "Task" [ 931.406825] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.416302] env[63355]: DEBUG oslo_vmware.api [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349771, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.430938] env[63355]: DEBUG nova.compute.manager [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 931.443330] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349768, 'name': ReconfigVM_Task, 'duration_secs': 0.227946} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.443643] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287631', 'volume_id': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'name': 'volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ecbacdb3-516e-43c3-96e3-4961b76565ca', 'attached_at': '', 'detached_at': '', 'volume_id': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'serial': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 931.444212] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4cd22c8-63ae-40bf-b0c5-2cf402a41090 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.452008] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for the task: (returnval){ [ 931.452008] env[63355]: value = "task-1349772" [ 931.452008] env[63355]: _type = "Task" [ 931.452008] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.460903] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349772, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.482360] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349769, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.576551] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49c2412d-a4f5-4fa5-ba0e-90f87a65adc4 tempest-AttachInterfacesV270Test-1582937945 tempest-AttachInterfacesV270Test-1582937945-project-member] Lock "b6056441-9ee4-484f-a1d2-077546f2c581" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.428s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.615165] env[63355]: DEBUG nova.network.neutron [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Successfully updated port: 43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 931.817961] env[63355]: DEBUG oslo_concurrency.lockutils [req-1ff212bf-8c26-4b4b-959e-a42714be7504 req-58cc832a-28ac-47b5-b3f2-8ed3212df05c service nova] Releasing lock "refresh_cache-d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.866117] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.866870] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.866870] env[63355]: DEBUG nova.network.neutron [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 931.881971] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5206922a-e4c2-b1d0-0879-34db2f3afd45, 'name': SearchDatastore_Task, 'duration_secs': 0.014415} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.882456] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-655880c6-5ed3-4ed5-9991-2cd89f696ffd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.888188] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for the task: (returnval){ [ 931.888188] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c0b42a-7864-9638-2170-7609d03eb980" [ 931.888188] env[63355]: _type = "Task" [ 931.888188] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.897414] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c0b42a-7864-9638-2170-7609d03eb980, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.919463] env[63355]: DEBUG oslo_vmware.api [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Task: {'id': task-1349771, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.501832} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.920849] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.921076] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 931.921271] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 931.921444] env[63355]: INFO nova.compute.manager [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Took 1.14 seconds to destroy the instance on the hypervisor. [ 931.921686] env[63355]: DEBUG oslo.service.loopingcall [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.923943] env[63355]: DEBUG nova.compute.manager [-] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.924079] env[63355]: DEBUG nova.network.neutron [-] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 931.943741] env[63355]: DEBUG nova.compute.manager [req-d82e03d6-4990-43e6-aa3b-a21e25e228bf req-a7eb243b-1b30-4f41-95ad-db760fc6fff8 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Received event network-changed-62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.943988] env[63355]: DEBUG nova.compute.manager [req-d82e03d6-4990-43e6-aa3b-a21e25e228bf req-a7eb243b-1b30-4f41-95ad-db760fc6fff8 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Refreshing instance network info cache due to event network-changed-62623c8d-eb2e-4e47-87f9-52f1df6a7d7f. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.944237] env[63355]: DEBUG oslo_concurrency.lockutils [req-d82e03d6-4990-43e6-aa3b-a21e25e228bf req-a7eb243b-1b30-4f41-95ad-db760fc6fff8 service nova] Acquiring lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.944391] env[63355]: DEBUG oslo_concurrency.lockutils [req-d82e03d6-4990-43e6-aa3b-a21e25e228bf req-a7eb243b-1b30-4f41-95ad-db760fc6fff8 service nova] Acquired lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.944558] env[63355]: DEBUG nova.network.neutron [req-d82e03d6-4990-43e6-aa3b-a21e25e228bf req-a7eb243b-1b30-4f41-95ad-db760fc6fff8 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Refreshing network info cache for port 62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.958788] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.963536] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349772, 'name': Rename_Task, 'duration_secs': 0.174781} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.963940] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.964218] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d820a241-7fa8-4dc0-9ad6-a7c38c1f6b69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.974721] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349769, 'name': CreateVM_Task, 'duration_secs': 0.554673} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.975842] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a408e19a-0960-430d-8550-0a304c63da61] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 931.976187] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for the task: (returnval){ [ 931.976187] env[63355]: value = "task-1349773" [ 931.976187] env[63355]: _type = "Task" [ 931.976187] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.976845] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.977032] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.977357] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 931.977653] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-102cf429-d01a-4ae1-b1d7-2bb5a9cfc77c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.988702] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349773, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.990049] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for the task: (returnval){ [ 931.990049] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524d1a99-d3bc-9f50-9910-2648969c7940" [ 931.990049] env[63355]: _type = "Task" [ 931.990049] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.000489] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524d1a99-d3bc-9f50-9910-2648969c7940, 'name': SearchDatastore_Task} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.001187] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.001187] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.001187] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.118618] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.118747] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.118862] env[63355]: DEBUG nova.network.neutron [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.395448] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a302b50-c67c-450b-8173-cad7dd41cb8a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.405409] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c0b42a-7864-9638-2170-7609d03eb980, 'name': SearchDatastore_Task, 'duration_secs': 0.016069} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.406137] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.406595] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] d2480bd6-527b-46b8-8a6a-22ad4eda3f5f/d2480bd6-527b-46b8-8a6a-22ad4eda3f5f.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 932.406950] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.407204] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.407684] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e4e6822-f6bc-4f68-a264-1ecd4a241485 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.411465] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19f6e561-3b1d-43bf-ba19-c9f8bf096c20 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.414213] env[63355]: WARNING nova.network.neutron [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] 943d3ecf-0c77-4c51-a997-b17d86259c34 already exists in list: networks containing: ['943d3ecf-0c77-4c51-a997-b17d86259c34']. ignoring it [ 932.416957] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bee916-a10d-46d1-8020-9866e108d423 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.421919] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for the task: (returnval){ [ 932.421919] env[63355]: value = "task-1349774" [ 932.421919] env[63355]: _type = "Task" [ 932.421919] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.456349] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.456349] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 932.462757] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a5adc88-65e4-4af4-aa26-f0c265e315a3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.466050] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466f718f-9087-445d-95ad-a43b39ec91ac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.469693] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349774, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.476725] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for the task: (returnval){ [ 932.476725] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52930e4b-5bdc-f2e3-5916-e129bda53623" [ 932.476725] env[63355]: _type = "Task" [ 932.476725] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.486929] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efa3747-1ee5-40d6-97ed-87dc60371eb7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.502845] env[63355]: DEBUG oslo_vmware.api [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349773, 'name': PowerOnVM_Task, 'duration_secs': 0.501036} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.516860] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.516860] env[63355]: INFO nova.compute.manager [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Took 8.37 seconds to spawn the instance on the hypervisor. [ 932.516860] env[63355]: DEBUG nova.compute.manager [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.516860] env[63355]: DEBUG nova.compute.provider_tree [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.516860] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52930e4b-5bdc-f2e3-5916-e129bda53623, 'name': SearchDatastore_Task, 'duration_secs': 0.015495} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.517878] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd95c412-3342-4f6d-b091-1a3311f32889 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.523102] env[63355]: INFO nova.network.neutron [req-d82e03d6-4990-43e6-aa3b-a21e25e228bf req-a7eb243b-1b30-4f41-95ad-db760fc6fff8 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Port 62623c8d-eb2e-4e47-87f9-52f1df6a7d7f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 932.523365] env[63355]: DEBUG nova.network.neutron [req-d82e03d6-4990-43e6-aa3b-a21e25e228bf req-a7eb243b-1b30-4f41-95ad-db760fc6fff8 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.525397] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86da7d0e-018e-4756-b07c-6a696a711901 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.537347] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for the task: (returnval){ [ 932.537347] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5252e2b2-7453-ed80-90c6-aaad966e0cd6" [ 932.537347] env[63355]: _type = "Task" [ 932.537347] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.548346] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5252e2b2-7453-ed80-90c6-aaad966e0cd6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.681738] env[63355]: DEBUG nova.network.neutron [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 932.717330] env[63355]: DEBUG nova.network.neutron [-] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.756569] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "dc56b350-cee7-49c4-9712-8c8c29146ff2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.756846] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "dc56b350-cee7-49c4-9712-8c8c29146ff2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.757070] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "dc56b350-cee7-49c4-9712-8c8c29146ff2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.757250] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "dc56b350-cee7-49c4-9712-8c8c29146ff2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.757417] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "dc56b350-cee7-49c4-9712-8c8c29146ff2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.760454] env[63355]: INFO nova.compute.manager [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Terminating instance [ 932.764922] env[63355]: DEBUG nova.network.neutron [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [{"id": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "address": "fa:16:3e:a9:13:18", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348dcf0-b9", "ovs_interfaceid": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "678307c8-eb2a-409d-a339-e40d5fd2ee50", "address": "fa:16:3e:9b:4c:95", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap678307c8-eb", "ovs_interfaceid": "678307c8-eb2a-409d-a339-e40d5fd2ee50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.769331] env[63355]: DEBUG nova.compute.manager [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 932.769331] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 932.771274] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e992f0-778b-4625-a2ed-dd541ccb0939 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.780471] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.782292] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-328f2811-8864-4b2c-83ac-9496bd0a810c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.789717] env[63355]: DEBUG oslo_vmware.api [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 932.789717] env[63355]: value = "task-1349775" [ 932.789717] env[63355]: _type = "Task" [ 932.789717] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.799690] env[63355]: DEBUG oslo_vmware.api [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349775, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.926481] env[63355]: DEBUG nova.compute.manager [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: a408e19a-0960-430d-8550-0a304c63da61] Received event network-changed-de9150b0-5d81-4397-a0d7-58c9a471d791 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.926818] env[63355]: DEBUG nova.compute.manager [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: a408e19a-0960-430d-8550-0a304c63da61] Refreshing instance network info cache due to event network-changed-de9150b0-5d81-4397-a0d7-58c9a471d791. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 932.927385] env[63355]: DEBUG oslo_concurrency.lockutils [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] Acquiring lock "refresh_cache-a408e19a-0960-430d-8550-0a304c63da61" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.927623] env[63355]: DEBUG oslo_concurrency.lockutils [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] Acquired lock "refresh_cache-a408e19a-0960-430d-8550-0a304c63da61" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.927923] env[63355]: DEBUG nova.network.neutron [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: a408e19a-0960-430d-8550-0a304c63da61] Refreshing network info cache for port de9150b0-5d81-4397-a0d7-58c9a471d791 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 932.941394] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349774, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.961239] env[63355]: DEBUG nova.compute.manager [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.962721] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0aa30a-3190-46ac-a65a-f47d8aa247ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.994133] env[63355]: DEBUG nova.network.neutron [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updating instance_info_cache with network_info: [{"id": "43ebe841-d4c8-446a-981c-519bef977228", "address": "fa:16:3e:97:62:7e", "network": {"id": "e19b4619-6e81-4205-838d-9ee7e7ae44d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1968447346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "744ce7dc0ab4463fa461cbac87e51904", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43ebe841-d4", "ovs_interfaceid": "43ebe841-d4c8-446a-981c-519bef977228", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.026800] env[63355]: DEBUG nova.scheduler.client.report [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.032185] env[63355]: DEBUG oslo_concurrency.lockutils [req-d82e03d6-4990-43e6-aa3b-a21e25e228bf req-a7eb243b-1b30-4f41-95ad-db760fc6fff8 service nova] Releasing lock "refresh_cache-74071e73-10fa-4dcb-aa15-91303b2278fb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.050806] env[63355]: INFO nova.compute.manager [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Took 28.90 seconds to build instance. [ 933.063088] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5252e2b2-7453-ed80-90c6-aaad966e0cd6, 'name': SearchDatastore_Task, 'duration_secs': 0.014204} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.063507] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.063507] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] a408e19a-0960-430d-8550-0a304c63da61/a408e19a-0960-430d-8550-0a304c63da61.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 933.066194] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd3c8c51-4d8b-48d9-8757-7d89b030d615 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.077106] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for the task: (returnval){ [ 933.077106] env[63355]: value = "task-1349776" [ 933.077106] env[63355]: _type = "Task" [ 933.077106] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.092522] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349776, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.221754] env[63355]: INFO nova.compute.manager [-] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Took 1.30 seconds to deallocate network for instance. [ 933.275525] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.276355] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.276575] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.277750] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6190e1ab-b741-40d3-997c-20285fffba33 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.299365] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.299632] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.299767] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.299972] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.300669] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.300851] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.301081] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.301250] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.301441] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.301624] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.301800] env[63355]: DEBUG nova.virt.hardware [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.308274] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Reconfiguring VM to attach interface {{(pid=63355) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 933.311892] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5486a53-52a0-48d3-bb2c-20fbd8d7c18d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.329919] env[63355]: DEBUG oslo_vmware.api [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349775, 'name': PowerOffVM_Task, 'duration_secs': 0.249449} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.331336] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.331520] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.332134] env[63355]: DEBUG oslo_vmware.api [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 933.332134] env[63355]: value = "task-1349777" [ 933.332134] env[63355]: _type = "Task" [ 933.332134] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.332364] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bdca544f-8edc-4121-acd7-971ed4c2a9ba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.342762] env[63355]: DEBUG oslo_vmware.api [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349777, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.409494] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.410023] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.410023] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Deleting the datastore file [datastore2] dc56b350-cee7-49c4-9712-8c8c29146ff2 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.410295] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6aa46d4c-7710-4825-820f-35a9bc3fcb17 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.417542] env[63355]: DEBUG oslo_vmware.api [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 933.417542] env[63355]: value = "task-1349779" [ 933.417542] env[63355]: _type = "Task" [ 933.417542] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.427882] env[63355]: DEBUG oslo_vmware.api [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.438945] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349774, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.863971} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.439274] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] d2480bd6-527b-46b8-8a6a-22ad4eda3f5f/d2480bd6-527b-46b8-8a6a-22ad4eda3f5f.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 933.439502] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 933.440339] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03055610-acbc-40bd-8ce8-e0463a2dc394 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.452902] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for the task: (returnval){ [ 933.452902] env[63355]: value = "task-1349780" [ 933.452902] env[63355]: _type = "Task" [ 933.452902] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.464796] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349780, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.481291] env[63355]: INFO nova.compute.manager [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] instance snapshotting [ 933.484416] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8480febe-089f-4051-9b2b-bdb577d564d0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.504111] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Releasing lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.506045] env[63355]: DEBUG nova.compute.manager [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Instance network_info: |[{"id": "43ebe841-d4c8-446a-981c-519bef977228", "address": "fa:16:3e:97:62:7e", "network": {"id": "e19b4619-6e81-4205-838d-9ee7e7ae44d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1968447346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "744ce7dc0ab4463fa461cbac87e51904", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43ebe841-d4", "ovs_interfaceid": "43ebe841-d4c8-446a-981c-519bef977228", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 933.507659] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:62:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16f09e8c-5240-4839-80cc-62ec29700bd2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43ebe841-d4c8-446a-981c-519bef977228', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.518092] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Creating folder: Project (744ce7dc0ab4463fa461cbac87e51904). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 933.518092] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434d806b-27f4-4136-85d6-711f8d296dad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.520689] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-231e33e1-e7d3-4d78-9bcd-b1e8c1c8249a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.536159] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.536159] env[63355]: DEBUG nova.compute.manager [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.537551] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Created folder: Project (744ce7dc0ab4463fa461cbac87e51904) in parent group-v287607. [ 933.541498] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Creating folder: Instances. Parent ref: group-v287693. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 933.541873] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.520s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.542123] env[63355]: DEBUG nova.objects.instance [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lazy-loading 'resources' on Instance uuid 2539a79e-01c0-4e0c-aa66-8784441c6fda {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.543182] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-512431c2-27fd-41ae-b814-9e94f4fd2e81 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.554354] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ce487cf0-1445-4190-bbb4-6b429158ccfc tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "ecbacdb3-516e-43c3-96e3-4961b76565ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.460s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.558200] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Created folder: Instances in parent group-v287693. [ 933.558200] env[63355]: DEBUG oslo.service.loopingcall [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.558373] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 933.558504] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a0fb7cb-6e84-47fd-8940-cf530858822c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.585441] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.585441] env[63355]: value = "task-1349783" [ 933.585441] env[63355]: _type = "Task" [ 933.585441] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.593342] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349776, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.601008] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349783, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.726941] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.765189] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.765659] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.766164] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.766380] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.766635] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.769815] env[63355]: INFO nova.compute.manager [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Terminating instance [ 933.772425] env[63355]: DEBUG nova.compute.manager [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 933.772660] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.773634] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f789c34-2370-42ae-baff-cc9e7466edb1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.787857] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.788958] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fbbb11f3-242c-4c24-baec-40e6cabeb474 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.797476] env[63355]: DEBUG oslo_vmware.api [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 933.797476] env[63355]: value = "task-1349784" [ 933.797476] env[63355]: _type = "Task" [ 933.797476] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.809452] env[63355]: DEBUG oslo_vmware.api [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.846320] env[63355]: DEBUG oslo_vmware.api [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349777, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.929385] env[63355]: DEBUG oslo_vmware.api [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349779, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.356418} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.929666] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.930452] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.930452] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.930452] env[63355]: INFO nova.compute.manager [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Took 1.16 seconds to destroy the instance on the hypervisor. [ 933.930602] env[63355]: DEBUG oslo.service.loopingcall [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.930723] env[63355]: DEBUG nova.compute.manager [-] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 933.930816] env[63355]: DEBUG nova.network.neutron [-] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 933.959208] env[63355]: DEBUG nova.network.neutron [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: a408e19a-0960-430d-8550-0a304c63da61] Updated VIF entry in instance network info cache for port de9150b0-5d81-4397-a0d7-58c9a471d791. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 933.959208] env[63355]: DEBUG nova.network.neutron [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: a408e19a-0960-430d-8550-0a304c63da61] Updating instance_info_cache with network_info: [{"id": "de9150b0-5d81-4397-a0d7-58c9a471d791", "address": "fa:16:3e:7d:8d:90", "network": {"id": "f0995df2-4470-43ed-8168-2d1324820fe0", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2109446375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c6bd8f13718c44e7bc41c7974ac90d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde9150b0-5d", "ovs_interfaceid": "de9150b0-5d81-4397-a0d7-58c9a471d791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.966576] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349780, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10377} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.966955] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.967808] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2420d99f-8e65-4b08-8b8a-ecf3d9ac326c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.992257] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] d2480bd6-527b-46b8-8a6a-22ad4eda3f5f/d2480bd6-527b-46b8-8a6a-22ad4eda3f5f.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.993235] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09eacd75-923f-40f0-a2db-2366c0c95410 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.018426] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for the task: (returnval){ [ 934.018426] env[63355]: value = "task-1349785" [ 934.018426] env[63355]: _type = "Task" [ 934.018426] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.030698] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349785, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.033149] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 934.033581] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8cbdf1b8-7f4d-4686-8ec4-313db32606aa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.044701] env[63355]: DEBUG nova.compute.utils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.046483] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 934.046483] env[63355]: value = "task-1349786" [ 934.046483] env[63355]: _type = "Task" [ 934.046483] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.047144] env[63355]: DEBUG nova.compute.manager [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.047383] env[63355]: DEBUG nova.network.neutron [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.063495] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349786, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.101832] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349776, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.830175} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.109821] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] a408e19a-0960-430d-8550-0a304c63da61/a408e19a-0960-430d-8550-0a304c63da61.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.109821] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.110127] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349783, 'name': CreateVM_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.110820] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f6122fe-b0e1-4f73-8b97-c71cb9561ab7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.120908] env[63355]: DEBUG nova.policy [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0f1d54b2d9c454688e7f56a40119657', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad0fba7800a641da9114e93e18d7977d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.128358] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for the task: (returnval){ [ 934.128358] env[63355]: value = "task-1349787" [ 934.128358] env[63355]: _type = "Task" [ 934.128358] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.138401] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349787, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.310421] env[63355]: DEBUG oslo_vmware.api [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349784, 'name': PowerOffVM_Task, 'duration_secs': 0.270637} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.310688] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.310850] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.311124] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd859764-bb2f-431d-9372-897d60521740 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.349063] env[63355]: DEBUG oslo_vmware.api [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349777, 'name': ReconfigVM_Task, 'duration_secs': 0.837155} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.349578] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.349810] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Reconfigured VM to attach interface {{(pid=63355) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 934.379184] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 934.379468] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 934.379666] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Deleting the datastore file [datastore2] 199bc488-2e5d-4cea-aefb-ddd35ecc7a30 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.379942] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94794ed9-b6bc-497a-9567-d0d4dbc91c78 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.391396] env[63355]: DEBUG oslo_vmware.api [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for the task: (returnval){ [ 934.391396] env[63355]: value = "task-1349789" [ 934.391396] env[63355]: _type = "Task" [ 934.391396] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.403386] env[63355]: DEBUG oslo_vmware.api [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349789, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.425477] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0a0b4d-06cd-4748-ae76-b8a8bd95a2f4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.433653] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a06950-93d2-4344-a6cf-6d42b24b3178 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.468559] env[63355]: DEBUG oslo_concurrency.lockutils [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] Releasing lock "refresh_cache-a408e19a-0960-430d-8550-0a304c63da61" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.468843] env[63355]: DEBUG nova.compute.manager [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received event network-vif-plugged-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.469060] env[63355]: DEBUG oslo_concurrency.lockutils [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] Acquiring lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.469283] env[63355]: DEBUG oslo_concurrency.lockutils [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.469447] env[63355]: DEBUG oslo_concurrency.lockutils [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.469619] env[63355]: DEBUG nova.compute.manager [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] No waiting events found dispatching network-vif-plugged-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.469804] env[63355]: WARNING nova.compute.manager [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received unexpected event network-vif-plugged-43ebe841-d4c8-446a-981c-519bef977228 for instance with vm_state building and task_state spawning. [ 934.470016] env[63355]: DEBUG nova.compute.manager [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received event network-changed-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.470187] env[63355]: DEBUG nova.compute.manager [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Refreshing instance network info cache due to event network-changed-43ebe841-d4c8-446a-981c-519bef977228. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.470378] env[63355]: DEBUG oslo_concurrency.lockutils [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] Acquiring lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.470561] env[63355]: DEBUG oslo_concurrency.lockutils [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] Acquired lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.470672] env[63355]: DEBUG nova.network.neutron [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Refreshing network info cache for port 43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 934.472915] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8b75e5-560f-461d-ae14-0f56217b974e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.482552] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f65f888-0184-41b1-9fff-7d52d3d2e82e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.499687] env[63355]: DEBUG nova.compute.provider_tree [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 934.528667] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349785, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.547746] env[63355]: DEBUG nova.compute.manager [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 934.561649] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349786, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.603434] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349783, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.642742] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349787, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.135515} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.643089] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.643889] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec75904-6854-43c6-aab5-06a75a5862b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.670364] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] a408e19a-0960-430d-8550-0a304c63da61/a408e19a-0960-430d-8550-0a304c63da61.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.670769] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55089160-874e-4504-9183-3877df06a03e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.694308] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for the task: (returnval){ [ 934.694308] env[63355]: value = "task-1349790" [ 934.694308] env[63355]: _type = "Task" [ 934.694308] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.703701] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349790, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.705593] env[63355]: DEBUG nova.network.neutron [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Successfully created port: 2f380960-ecf1-407f-aaf0-88703f8132c3 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.755128] env[63355]: DEBUG nova.compute.manager [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received event network-vif-plugged-678307c8-eb2a-409d-a339-e40d5fd2ee50 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.755128] env[63355]: DEBUG oslo_concurrency.lockutils [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] Acquiring lock "386e847e-967b-4247-9730-cdc5ac251474-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.755128] env[63355]: DEBUG oslo_concurrency.lockutils [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] Lock "386e847e-967b-4247-9730-cdc5ac251474-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.755851] env[63355]: DEBUG oslo_concurrency.lockutils [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] Lock "386e847e-967b-4247-9730-cdc5ac251474-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.755851] env[63355]: DEBUG nova.compute.manager [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] No waiting events found dispatching network-vif-plugged-678307c8-eb2a-409d-a339-e40d5fd2ee50 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.755851] env[63355]: WARNING nova.compute.manager [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received unexpected event network-vif-plugged-678307c8-eb2a-409d-a339-e40d5fd2ee50 for instance with vm_state active and task_state None. [ 934.756035] env[63355]: DEBUG nova.compute.manager [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received event network-changed-678307c8-eb2a-409d-a339-e40d5fd2ee50 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.756158] env[63355]: DEBUG nova.compute.manager [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Refreshing instance network info cache due to event network-changed-678307c8-eb2a-409d-a339-e40d5fd2ee50. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.756356] env[63355]: DEBUG oslo_concurrency.lockutils [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] Acquiring lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.756556] env[63355]: DEBUG oslo_concurrency.lockutils [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] Acquired lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.756750] env[63355]: DEBUG nova.network.neutron [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Refreshing network info cache for port 678307c8-eb2a-409d-a339-e40d5fd2ee50 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 934.784683] env[63355]: DEBUG nova.network.neutron [-] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.854653] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0a4850c7-0f36-4778-aca2-78ee743482e9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-386e847e-967b-4247-9730-cdc5ac251474-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.018s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.901435] env[63355]: DEBUG oslo_vmware.api [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Task: {'id': task-1349789, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17306} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.901834] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.901929] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.902179] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.902317] env[63355]: INFO nova.compute.manager [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Took 1.13 seconds to destroy the instance on the hypervisor. [ 934.902799] env[63355]: DEBUG oslo.service.loopingcall [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.902869] env[63355]: DEBUG nova.compute.manager [-] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 934.902930] env[63355]: DEBUG nova.network.neutron [-] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 935.031611] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349785, 'name': ReconfigVM_Task, 'duration_secs': 0.632726} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.034379] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Reconfigured VM instance instance-0000004a to attach disk [datastore1] d2480bd6-527b-46b8-8a6a-22ad4eda3f5f/d2480bd6-527b-46b8-8a6a-22ad4eda3f5f.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.035146] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ca111596-bdcc-41e1-a446-fce82b1a2aba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.045945] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for the task: (returnval){ [ 935.045945] env[63355]: value = "task-1349791" [ 935.045945] env[63355]: _type = "Task" [ 935.045945] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.060588] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349791, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.061956] env[63355]: DEBUG nova.scheduler.client.report [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Updated inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b with generation 93 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 935.062194] env[63355]: DEBUG nova.compute.provider_tree [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Updating resource provider 47abb610-db7e-4770-911d-187dd075ef8b generation from 93 to 94 during operation: update_inventory {{(pid=63355) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 935.062339] env[63355]: DEBUG nova.compute.provider_tree [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.073204] env[63355]: DEBUG nova.network.neutron [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Successfully created port: cf363065-5623-4cf2-9d11-fcc3a234909a {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 935.084514] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349786, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.104513] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349783, 'name': CreateVM_Task, 'duration_secs': 1.098111} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.105628] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 935.106179] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.106605] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.108123] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 935.108123] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdc3ac62-e753-439c-a184-e39b830fd8f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.115700] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 935.115700] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52329af8-fb7b-6476-cb1e-1bb1f51aec17" [ 935.115700] env[63355]: _type = "Task" [ 935.115700] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.132916] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52329af8-fb7b-6476-cb1e-1bb1f51aec17, 'name': SearchDatastore_Task, 'duration_secs': 0.011881} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.132916] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.133218] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.133502] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.133674] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.134191] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.134278] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dbd46567-0b5b-464d-a5cf-7cbaebd10249 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.144746] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.144937] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 935.146075] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5f9824d-0857-420e-ab15-240136d75564 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.154965] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 935.154965] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52485ce6-0005-dd71-292c-ef5e4220a623" [ 935.154965] env[63355]: _type = "Task" [ 935.154965] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.167965] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52485ce6-0005-dd71-292c-ef5e4220a623, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.208901] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349790, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.287256] env[63355]: INFO nova.compute.manager [-] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Took 1.36 seconds to deallocate network for instance. [ 935.358321] env[63355]: DEBUG nova.network.neutron [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updated VIF entry in instance network info cache for port 43ebe841-d4c8-446a-981c-519bef977228. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 935.359026] env[63355]: DEBUG nova.network.neutron [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updating instance_info_cache with network_info: [{"id": "43ebe841-d4c8-446a-981c-519bef977228", "address": "fa:16:3e:97:62:7e", "network": {"id": "e19b4619-6e81-4205-838d-9ee7e7ae44d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1968447346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "744ce7dc0ab4463fa461cbac87e51904", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43ebe841-d4", "ovs_interfaceid": "43ebe841-d4c8-446a-981c-519bef977228", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.564525] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349791, 'name': Rename_Task, 'duration_secs': 0.227275} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.570108] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 935.571310] env[63355]: DEBUG nova.compute.manager [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.575195] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e884811-e4ef-443d-951b-7a92fa0077cc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.580570] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.039s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.588544] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.530s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.589133] env[63355]: DEBUG nova.objects.instance [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63355) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 935.596191] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349786, 'name': CreateSnapshot_Task, 'duration_secs': 1.147669} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.596955] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 935.601153] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5321529-64aa-481d-9a11-d2305f83d020 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.603147] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for the task: (returnval){ [ 935.603147] env[63355]: value = "task-1349792" [ 935.603147] env[63355]: _type = "Task" [ 935.603147] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.615656] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.615975] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.616172] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.616365] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.616515] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.616713] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.617018] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.617205] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.617379] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.617548] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.617721] env[63355]: DEBUG nova.virt.hardware [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.622978] env[63355]: INFO nova.scheduler.client.report [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Deleted allocations for instance 2539a79e-01c0-4e0c-aa66-8784441c6fda [ 935.624585] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c64bde-6750-407b-8287-79c8b226376f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.648037] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07414c0f-5c33-48ca-959a-0faa92513eec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.652212] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349792, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.672474] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52485ce6-0005-dd71-292c-ef5e4220a623, 'name': SearchDatastore_Task, 'duration_secs': 0.012233} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.673307] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b1e9e10-f76d-4ec6-b962-38412a584344 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.679516] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 935.679516] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5286a678-83b3-06b3-7047-65b701311cb6" [ 935.679516] env[63355]: _type = "Task" [ 935.679516] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.688425] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5286a678-83b3-06b3-7047-65b701311cb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.704511] env[63355]: DEBUG nova.compute.manager [req-d928e5ff-a8d4-4d70-aec0-0665db748f8b req-1bb67bd2-fbd5-4ffb-b1b6-ac1c41244a66 service nova] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Received event network-vif-deleted-62623c8d-eb2e-4e47-87f9-52f1df6a7d7f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.704713] env[63355]: DEBUG nova.compute.manager [req-d928e5ff-a8d4-4d70-aec0-0665db748f8b req-1bb67bd2-fbd5-4ffb-b1b6-ac1c41244a66 service nova] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Received event network-vif-deleted-42e50136-f618-4d11-82e9-5efca796fd5b {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.711589] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349790, 'name': ReconfigVM_Task, 'duration_secs': 0.551799} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.711860] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Reconfigured VM instance instance-0000004b to attach disk [datastore1] a408e19a-0960-430d-8550-0a304c63da61/a408e19a-0960-430d-8550-0a304c63da61.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.712712] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4dd2b9e5-bf09-47f1-be00-d866835847e2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.720592] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for the task: (returnval){ [ 935.720592] env[63355]: value = "task-1349793" [ 935.720592] env[63355]: _type = "Task" [ 935.720592] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.731914] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349793, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.794186] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.836880] env[63355]: DEBUG nova.network.neutron [-] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.863838] env[63355]: DEBUG oslo_concurrency.lockutils [req-9034befd-d2fd-4aab-810f-2fd6326103da req-c933106b-d166-4bc3-9285-3cd6a277181d service nova] Releasing lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.864238] env[63355]: DEBUG nova.network.neutron [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updated VIF entry in instance network info cache for port 678307c8-eb2a-409d-a339-e40d5fd2ee50. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 935.864661] env[63355]: DEBUG nova.network.neutron [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [{"id": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "address": "fa:16:3e:a9:13:18", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348dcf0-b9", "ovs_interfaceid": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "678307c8-eb2a-409d-a339-e40d5fd2ee50", "address": "fa:16:3e:9b:4c:95", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap678307c8-eb", "ovs_interfaceid": "678307c8-eb2a-409d-a339-e40d5fd2ee50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.117898] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349792, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.147418] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 936.148558] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a1fcc28c-b3d7-4264-b0d5-490e0ed6bca0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.152542] env[63355]: DEBUG oslo_concurrency.lockutils [None req-26d84900-1044-4dde-b8e5-27dfa3dc5153 tempest-ServerAddressesTestJSON-330334585 tempest-ServerAddressesTestJSON-330334585-project-member] Lock "2539a79e-01c0-4e0c-aa66-8784441c6fda" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.267s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.160766] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 936.160766] env[63355]: value = "task-1349794" [ 936.160766] env[63355]: _type = "Task" [ 936.160766] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.170153] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349794, 'name': CloneVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.190130] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5286a678-83b3-06b3-7047-65b701311cb6, 'name': SearchDatastore_Task, 'duration_secs': 0.011785} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.190451] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.190710] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 7c153109-b814-4e11-b4f9-7b8cebb853d4/7c153109-b814-4e11-b4f9-7b8cebb853d4.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 936.190984] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cad35f38-8f86-4281-9ad1-bff9d2b9a83f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.198526] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 936.198526] env[63355]: value = "task-1349795" [ 936.198526] env[63355]: _type = "Task" [ 936.198526] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.207388] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349795, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.231213] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349793, 'name': Rename_Task, 'duration_secs': 0.3468} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.231521] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.231786] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aca48059-0950-4b18-b8f5-7c6410e00ba2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.239330] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for the task: (returnval){ [ 936.239330] env[63355]: value = "task-1349796" [ 936.239330] env[63355]: _type = "Task" [ 936.239330] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.248415] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349796, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.339465] env[63355]: INFO nova.compute.manager [-] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Took 1.44 seconds to deallocate network for instance. [ 936.367467] env[63355]: DEBUG oslo_concurrency.lockutils [req-fac8b3bf-dfec-4a59-918d-01b7a83723c6 req-823e8cd9-c24e-4d65-bbe0-1ad4d2672778 service nova] Releasing lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.603072] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8766a2f-ffad-46be-9dbc-cfb009864c31 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.604474] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.714s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.604533] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.604843] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 936.605411] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.936s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.607944] env[63355]: INFO nova.compute.claims [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.612850] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f133575e-a45b-444b-bfe1-7b6607d8d393 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.631417] env[63355]: DEBUG oslo_vmware.api [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349792, 'name': PowerOnVM_Task, 'duration_secs': 1.021961} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.635288] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 936.635811] env[63355]: INFO nova.compute.manager [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Took 10.67 seconds to spawn the instance on the hypervisor. [ 936.636067] env[63355]: DEBUG nova.compute.manager [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.638781] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ed50c8-c627-4575-b31d-c935d813c923 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.643033] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd4596a-60f0-4a73-9ea2-780dd89e757e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.679381] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1cbea7a-733b-4f4b-bb5c-932b878d9038 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.697584] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd27bb3-31b3-4757-8c4f-48f6e58d6a43 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.701900] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349794, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.712968] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349795, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.743129] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180874MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 936.743129] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.753799] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349796, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.846289] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.136706] env[63355]: DEBUG nova.compute.manager [req-099e49bd-c9a4-4164-ba2e-503f3f636026 req-86d620c8-eeea-46c1-a3c1-5686a0835de2 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Received event network-vif-plugged-2f380960-ecf1-407f-aaf0-88703f8132c3 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.137015] env[63355]: DEBUG oslo_concurrency.lockutils [req-099e49bd-c9a4-4164-ba2e-503f3f636026 req-86d620c8-eeea-46c1-a3c1-5686a0835de2 service nova] Acquiring lock "6706c65c-3959-471b-82c0-2196f4aa1ab0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.137194] env[63355]: DEBUG oslo_concurrency.lockutils [req-099e49bd-c9a4-4164-ba2e-503f3f636026 req-86d620c8-eeea-46c1-a3c1-5686a0835de2 service nova] Lock "6706c65c-3959-471b-82c0-2196f4aa1ab0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.137314] env[63355]: DEBUG oslo_concurrency.lockutils [req-099e49bd-c9a4-4164-ba2e-503f3f636026 req-86d620c8-eeea-46c1-a3c1-5686a0835de2 service nova] Lock "6706c65c-3959-471b-82c0-2196f4aa1ab0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.137484] env[63355]: DEBUG nova.compute.manager [req-099e49bd-c9a4-4164-ba2e-503f3f636026 req-86d620c8-eeea-46c1-a3c1-5686a0835de2 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] No waiting events found dispatching network-vif-plugged-2f380960-ecf1-407f-aaf0-88703f8132c3 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.137649] env[63355]: WARNING nova.compute.manager [req-099e49bd-c9a4-4164-ba2e-503f3f636026 req-86d620c8-eeea-46c1-a3c1-5686a0835de2 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Received unexpected event network-vif-plugged-2f380960-ecf1-407f-aaf0-88703f8132c3 for instance with vm_state building and task_state spawning. [ 937.176900] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349794, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.200373] env[63355]: INFO nova.compute.manager [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Took 32.74 seconds to build instance. [ 937.213971] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349795, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579135} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.214268] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 7c153109-b814-4e11-b4f9-7b8cebb853d4/7c153109-b814-4e11-b4f9-7b8cebb853d4.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 937.214602] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 937.214732] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-666ba5d1-e191-47d0-afdb-0c35bb2bed38 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.222846] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 937.222846] env[63355]: value = "task-1349797" [ 937.222846] env[63355]: _type = "Task" [ 937.222846] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.234985] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349797, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.252544] env[63355]: DEBUG oslo_vmware.api [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349796, 'name': PowerOnVM_Task, 'duration_secs': 0.611962} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.252946] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.253066] env[63355]: INFO nova.compute.manager [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Took 8.72 seconds to spawn the instance on the hypervisor. [ 937.253237] env[63355]: DEBUG nova.compute.manager [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.254152] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e355a0-5f4f-487a-a553-fd4485c2f6cb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.318554] env[63355]: DEBUG nova.network.neutron [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Successfully updated port: 2f380960-ecf1-407f-aaf0-88703f8132c3 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.679446] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349794, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.704014] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9433ff9f-6981-4b70-a6bf-2d473841250a tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.124s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.704417] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "interface-386e847e-967b-4247-9730-cdc5ac251474-bd4c85c2-7841-4da0-ba95-43109bfd107e" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.704643] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-386e847e-967b-4247-9730-cdc5ac251474-bd4c85c2-7841-4da0-ba95-43109bfd107e" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.705224] env[63355]: DEBUG nova.objects.instance [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'flavor' on Instance uuid 386e847e-967b-4247-9730-cdc5ac251474 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.736132] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349797, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065678} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.736467] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.738375] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61a77d9-a379-42da-ab90-d516b4b37c77 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.742599] env[63355]: DEBUG nova.compute.manager [req-adba92ce-f902-4e6e-ac03-c0b1a7313820 req-4f8e0a0b-f074-419d-80ff-a0ef79345555 service nova] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Received event network-vif-deleted-b6165b88-87f9-4c10-bc5c-fc58b48d167c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.742810] env[63355]: DEBUG nova.compute.manager [req-adba92ce-f902-4e6e-ac03-c0b1a7313820 req-4f8e0a0b-f074-419d-80ff-a0ef79345555 service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Received event network-changed-0500cf25-c654-4748-a58a-8e51bd919768 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.742977] env[63355]: DEBUG nova.compute.manager [req-adba92ce-f902-4e6e-ac03-c0b1a7313820 req-4f8e0a0b-f074-419d-80ff-a0ef79345555 service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Refreshing instance network info cache due to event network-changed-0500cf25-c654-4748-a58a-8e51bd919768. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.743204] env[63355]: DEBUG oslo_concurrency.lockutils [req-adba92ce-f902-4e6e-ac03-c0b1a7313820 req-4f8e0a0b-f074-419d-80ff-a0ef79345555 service nova] Acquiring lock "refresh_cache-ecbacdb3-516e-43c3-96e3-4961b76565ca" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.743347] env[63355]: DEBUG oslo_concurrency.lockutils [req-adba92ce-f902-4e6e-ac03-c0b1a7313820 req-4f8e0a0b-f074-419d-80ff-a0ef79345555 service nova] Acquired lock "refresh_cache-ecbacdb3-516e-43c3-96e3-4961b76565ca" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.743507] env[63355]: DEBUG nova.network.neutron [req-adba92ce-f902-4e6e-ac03-c0b1a7313820 req-4f8e0a0b-f074-419d-80ff-a0ef79345555 service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Refreshing network info cache for port 0500cf25-c654-4748-a58a-8e51bd919768 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.770152] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 7c153109-b814-4e11-b4f9-7b8cebb853d4/7c153109-b814-4e11-b4f9-7b8cebb853d4.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.779246] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ea61611-c976-46ab-bba2-264c343289ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.794398] env[63355]: INFO nova.compute.manager [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Took 29.52 seconds to build instance. [ 937.802362] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 937.802362] env[63355]: value = "task-1349798" [ 937.802362] env[63355]: _type = "Task" [ 937.802362] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.825446] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349798, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.048377] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802accd8-4697-4a7f-96d5-5be006e5d9be {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.058250] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854d2a32-54f6-44eb-8e55-0446ffc88e45 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.098197] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a84b7ca-8654-40c5-bbc0-f9b434be3439 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.110181] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75106a32-9bde-4e11-b766-299f4229e674 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.127420] env[63355]: DEBUG nova.compute.provider_tree [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.180215] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349794, 'name': CloneVM_Task, 'duration_secs': 1.897849} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.180537] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Created linked-clone VM from snapshot [ 938.181362] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b7ccc1-9ee9-45d5-8f77-1e3ccc9b0d9a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.191428] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Uploading image faa91285-130f-41be-b93d-fe59ddde3549 {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 938.220472] env[63355]: DEBUG oslo_vmware.rw_handles [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 938.220472] env[63355]: value = "vm-287697" [ 938.220472] env[63355]: _type = "VirtualMachine" [ 938.220472] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 938.220765] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8af484e8-c5cb-47b2-b21a-19b7ad5eff60 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.230692] env[63355]: DEBUG oslo_vmware.rw_handles [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lease: (returnval){ [ 938.230692] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b572e1-f1d4-dd30-6139-6720e781526d" [ 938.230692] env[63355]: _type = "HttpNfcLease" [ 938.230692] env[63355]: } obtained for exporting VM: (result){ [ 938.230692] env[63355]: value = "vm-287697" [ 938.230692] env[63355]: _type = "VirtualMachine" [ 938.230692] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 938.231088] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the lease: (returnval){ [ 938.231088] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b572e1-f1d4-dd30-6139-6720e781526d" [ 938.231088] env[63355]: _type = "HttpNfcLease" [ 938.231088] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 938.238846] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 938.238846] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b572e1-f1d4-dd30-6139-6720e781526d" [ 938.238846] env[63355]: _type = "HttpNfcLease" [ 938.238846] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 938.296757] env[63355]: DEBUG oslo_concurrency.lockutils [None req-96245654-973b-4bab-858c-4d503b636644 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Lock "a408e19a-0960-430d-8550-0a304c63da61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.684s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.313694] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349798, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.439466] env[63355]: DEBUG nova.objects.instance [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'pci_requests' on Instance uuid 386e847e-967b-4247-9730-cdc5ac251474 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.518405] env[63355]: DEBUG nova.network.neutron [req-adba92ce-f902-4e6e-ac03-c0b1a7313820 req-4f8e0a0b-f074-419d-80ff-a0ef79345555 service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Updated VIF entry in instance network info cache for port 0500cf25-c654-4748-a58a-8e51bd919768. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 938.518821] env[63355]: DEBUG nova.network.neutron [req-adba92ce-f902-4e6e-ac03-c0b1a7313820 req-4f8e0a0b-f074-419d-80ff-a0ef79345555 service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Updating instance_info_cache with network_info: [{"id": "0500cf25-c654-4748-a58a-8e51bd919768", "address": "fa:16:3e:68:a5:c4", "network": {"id": "cca3f41d-524d-4176-871b-a2e826353fed", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1431855559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c787ea7464c54aa9a443b4945e8cb5df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0500cf25-c6", "ovs_interfaceid": "0500cf25-c654-4748-a58a-8e51bd919768", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.631854] env[63355]: DEBUG nova.scheduler.client.report [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.741242] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 938.741242] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b572e1-f1d4-dd30-6139-6720e781526d" [ 938.741242] env[63355]: _type = "HttpNfcLease" [ 938.741242] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 938.741450] env[63355]: DEBUG oslo_vmware.rw_handles [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 938.741450] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b572e1-f1d4-dd30-6139-6720e781526d" [ 938.741450] env[63355]: _type = "HttpNfcLease" [ 938.741450] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 938.742058] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468842ea-d623-43cc-b389-17d186eca243 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.752653] env[63355]: DEBUG oslo_vmware.rw_handles [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ba933-dea8-3806-76b5-deed25ade1ba/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 938.752912] env[63355]: DEBUG oslo_vmware.rw_handles [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ba933-dea8-3806-76b5-deed25ade1ba/disk-0.vmdk for reading. {{(pid=63355) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 938.827051] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349798, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.870892] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b40ed75e-de0f-4ecf-b33c-84ea37fdd6e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.943557] env[63355]: DEBUG nova.objects.base [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Object Instance<386e847e-967b-4247-9730-cdc5ac251474> lazy-loaded attributes: flavor,pci_requests {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 938.943994] env[63355]: DEBUG nova.network.neutron [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 939.021668] env[63355]: DEBUG oslo_concurrency.lockutils [req-adba92ce-f902-4e6e-ac03-c0b1a7313820 req-4f8e0a0b-f074-419d-80ff-a0ef79345555 service nova] Releasing lock "refresh_cache-ecbacdb3-516e-43c3-96e3-4961b76565ca" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.042981] env[63355]: DEBUG nova.policy [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77243643ea724b72858a8682a2a054a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d47b671ea9c429391cbdae7e24adadf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 939.136512] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.531s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.137157] env[63355]: DEBUG nova.compute.manager [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 939.140327] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.710s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.142069] env[63355]: INFO nova.compute.claims [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.261140] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquiring lock "a408e19a-0960-430d-8550-0a304c63da61" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.261140] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Lock "a408e19a-0960-430d-8550-0a304c63da61" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.264081] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquiring lock "a408e19a-0960-430d-8550-0a304c63da61-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.264081] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Lock "a408e19a-0960-430d-8550-0a304c63da61-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.264081] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Lock "a408e19a-0960-430d-8550-0a304c63da61-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.266511] env[63355]: INFO nova.compute.manager [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Terminating instance [ 939.269428] env[63355]: DEBUG nova.compute.manager [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 939.269882] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.273902] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed88217-5bb5-438d-bb70-e72830118062 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.280536] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.281069] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d8caf8c-6fdb-4142-b59c-d761fe99dca0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.290551] env[63355]: DEBUG oslo_vmware.api [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for the task: (returnval){ [ 939.290551] env[63355]: value = "task-1349800" [ 939.290551] env[63355]: _type = "Task" [ 939.290551] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.303501] env[63355]: DEBUG oslo_vmware.api [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349800, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.325700] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349798, 'name': ReconfigVM_Task, 'duration_secs': 1.475988} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.326149] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 7c153109-b814-4e11-b4f9-7b8cebb853d4/7c153109-b814-4e11-b4f9-7b8cebb853d4.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.327078] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ca32e96f-d9ba-4be2-ad2f-d77d1ce5ecbf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.338021] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 939.338021] env[63355]: value = "task-1349801" [ 939.338021] env[63355]: _type = "Task" [ 939.338021] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.347478] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349801, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.655330] env[63355]: DEBUG nova.compute.utils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 939.659790] env[63355]: DEBUG nova.compute.manager [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 939.659926] env[63355]: DEBUG nova.network.neutron [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 939.664212] env[63355]: DEBUG nova.compute.manager [req-ee95781f-e6db-4282-ba30-634ea4e7aecd req-dfe58f56-5bcf-4bc2-a940-9e2f992aeb47 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Received event network-changed-2f380960-ecf1-407f-aaf0-88703f8132c3 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.664212] env[63355]: DEBUG nova.compute.manager [req-ee95781f-e6db-4282-ba30-634ea4e7aecd req-dfe58f56-5bcf-4bc2-a940-9e2f992aeb47 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Refreshing instance network info cache due to event network-changed-2f380960-ecf1-407f-aaf0-88703f8132c3. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 939.664212] env[63355]: DEBUG oslo_concurrency.lockutils [req-ee95781f-e6db-4282-ba30-634ea4e7aecd req-dfe58f56-5bcf-4bc2-a940-9e2f992aeb47 service nova] Acquiring lock "refresh_cache-6706c65c-3959-471b-82c0-2196f4aa1ab0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.664212] env[63355]: DEBUG oslo_concurrency.lockutils [req-ee95781f-e6db-4282-ba30-634ea4e7aecd req-dfe58f56-5bcf-4bc2-a940-9e2f992aeb47 service nova] Acquired lock "refresh_cache-6706c65c-3959-471b-82c0-2196f4aa1ab0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.664212] env[63355]: DEBUG nova.network.neutron [req-ee95781f-e6db-4282-ba30-634ea4e7aecd req-dfe58f56-5bcf-4bc2-a940-9e2f992aeb47 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Refreshing network info cache for port 2f380960-ecf1-407f-aaf0-88703f8132c3 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 939.734240] env[63355]: DEBUG nova.policy [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'beb41983e69a44b2a108f552da5ef640', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfff1b9903264e5586119ebd3a3602de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 939.803980] env[63355]: DEBUG oslo_vmware.api [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349800, 'name': PowerOffVM_Task, 'duration_secs': 0.289545} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.805057] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.805057] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.806016] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6c1b223-5455-44b7-a773-d7f99c59f720 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.847382] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349801, 'name': Rename_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.883808] env[63355]: DEBUG nova.network.neutron [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Successfully updated port: cf363065-5623-4cf2-9d11-fcc3a234909a {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 939.892777] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.895106] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.895106] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Deleting the datastore file [datastore1] a408e19a-0960-430d-8550-0a304c63da61 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.895106] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be946aff-f1d5-4fcc-9075-978b284c4f74 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.905357] env[63355]: DEBUG oslo_vmware.api [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for the task: (returnval){ [ 939.905357] env[63355]: value = "task-1349803" [ 939.905357] env[63355]: _type = "Task" [ 939.905357] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.918268] env[63355]: DEBUG oslo_vmware.api [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.148641] env[63355]: DEBUG nova.network.neutron [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Successfully created port: 850ef0bb-ad5f-4627-b482-a78817b84343 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 940.164185] env[63355]: DEBUG nova.compute.manager [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 940.241390] env[63355]: DEBUG nova.network.neutron [req-ee95781f-e6db-4282-ba30-634ea4e7aecd req-dfe58f56-5bcf-4bc2-a940-9e2f992aeb47 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 940.349587] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349801, 'name': Rename_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.389699] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "refresh_cache-6706c65c-3959-471b-82c0-2196f4aa1ab0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.424570] env[63355]: DEBUG oslo_vmware.api [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Task: {'id': task-1349803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135388} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.425000] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.425311] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.425893] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.425893] env[63355]: INFO nova.compute.manager [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] [instance: a408e19a-0960-430d-8550-0a304c63da61] Took 1.16 seconds to destroy the instance on the hypervisor. [ 940.425999] env[63355]: DEBUG oslo.service.loopingcall [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.426264] env[63355]: DEBUG nova.compute.manager [-] [instance: a408e19a-0960-430d-8550-0a304c63da61] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.429862] env[63355]: DEBUG nova.network.neutron [-] [instance: a408e19a-0960-430d-8550-0a304c63da61] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 940.444914] env[63355]: DEBUG nova.network.neutron [req-ee95781f-e6db-4282-ba30-634ea4e7aecd req-dfe58f56-5bcf-4bc2-a940-9e2f992aeb47 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.505260] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068dc187-8094-4c26-873c-a3e99e2e6f81 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.514832] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bfe2baa-bac2-4ebb-b586-5443c9a3be05 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.549317] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d85765-988a-4f41-a8cf-bf0e2305a99c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.559982] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce533f3a-618e-49c9-ba7e-70209eb4e079 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.577565] env[63355]: DEBUG nova.compute.provider_tree [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.853942] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349801, 'name': Rename_Task, 'duration_secs': 1.170988} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.854462] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 940.855097] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e04cca5f-90f7-4000-8ee8-1ba190027cb4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.865120] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 940.865120] env[63355]: value = "task-1349804" [ 940.865120] env[63355]: _type = "Task" [ 940.865120] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.878204] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349804, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.951567] env[63355]: DEBUG oslo_concurrency.lockutils [req-ee95781f-e6db-4282-ba30-634ea4e7aecd req-dfe58f56-5bcf-4bc2-a940-9e2f992aeb47 service nova] Releasing lock "refresh_cache-6706c65c-3959-471b-82c0-2196f4aa1ab0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.951567] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquired lock "refresh_cache-6706c65c-3959-471b-82c0-2196f4aa1ab0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.951567] env[63355]: DEBUG nova.network.neutron [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.071182] env[63355]: DEBUG nova.network.neutron [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Successfully updated port: bd4c85c2-7841-4da0-ba95-43109bfd107e {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.084025] env[63355]: DEBUG nova.scheduler.client.report [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.101589] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.101754] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.102050] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.102339] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.102602] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.106349] env[63355]: INFO nova.compute.manager [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Terminating instance [ 941.111616] env[63355]: DEBUG nova.compute.manager [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 941.112395] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 941.113485] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac79294-282a-4f23-bf0b-78f38e58a134 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.125300] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 941.125736] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d1a95fb-8b93-4382-abae-b6b3d2ea265e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.137499] env[63355]: DEBUG oslo_vmware.api [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for the task: (returnval){ [ 941.137499] env[63355]: value = "task-1349805" [ 941.137499] env[63355]: _type = "Task" [ 941.137499] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.145771] env[63355]: DEBUG oslo_vmware.api [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349805, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.178577] env[63355]: DEBUG nova.compute.manager [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 941.226997] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.227342] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.227522] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.227892] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.228232] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.228427] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.228667] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.228848] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.229058] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.229274] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.229439] env[63355]: DEBUG nova.virt.hardware [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.230419] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed243e63-bff8-45a8-943e-0311e4009fbb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.242219] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64294650-1968-4be9-a573-60b02d7e02a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.332407] env[63355]: DEBUG nova.compute.manager [req-2ca3b2ae-168d-4781-ba45-06a0c2006d9e req-6faaef8a-c4d3-43cc-a69b-826caa81a685 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received event network-vif-plugged-bd4c85c2-7841-4da0-ba95-43109bfd107e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.332633] env[63355]: DEBUG oslo_concurrency.lockutils [req-2ca3b2ae-168d-4781-ba45-06a0c2006d9e req-6faaef8a-c4d3-43cc-a69b-826caa81a685 service nova] Acquiring lock "386e847e-967b-4247-9730-cdc5ac251474-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.333293] env[63355]: DEBUG oslo_concurrency.lockutils [req-2ca3b2ae-168d-4781-ba45-06a0c2006d9e req-6faaef8a-c4d3-43cc-a69b-826caa81a685 service nova] Lock "386e847e-967b-4247-9730-cdc5ac251474-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.333783] env[63355]: DEBUG oslo_concurrency.lockutils [req-2ca3b2ae-168d-4781-ba45-06a0c2006d9e req-6faaef8a-c4d3-43cc-a69b-826caa81a685 service nova] Lock "386e847e-967b-4247-9730-cdc5ac251474-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.334039] env[63355]: DEBUG nova.compute.manager [req-2ca3b2ae-168d-4781-ba45-06a0c2006d9e req-6faaef8a-c4d3-43cc-a69b-826caa81a685 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] No waiting events found dispatching network-vif-plugged-bd4c85c2-7841-4da0-ba95-43109bfd107e {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.334248] env[63355]: WARNING nova.compute.manager [req-2ca3b2ae-168d-4781-ba45-06a0c2006d9e req-6faaef8a-c4d3-43cc-a69b-826caa81a685 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received unexpected event network-vif-plugged-bd4c85c2-7841-4da0-ba95-43109bfd107e for instance with vm_state active and task_state None. [ 941.377490] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349804, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.500400] env[63355]: DEBUG nova.network.neutron [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 941.554175] env[63355]: DEBUG nova.network.neutron [-] [instance: a408e19a-0960-430d-8550-0a304c63da61] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.574527] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.574735] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.574916] env[63355]: DEBUG nova.network.neutron [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.587871] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.588467] env[63355]: DEBUG nova.compute.manager [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 941.591357] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.099s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.593786] env[63355]: DEBUG nova.objects.instance [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lazy-loading 'resources' on Instance uuid 7c139710-d8d9-4cd7-bec0-6e021d3b2e68 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.651267] env[63355]: DEBUG oslo_vmware.api [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349805, 'name': PowerOffVM_Task, 'duration_secs': 0.281673} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.651909] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.652179] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 941.652687] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22242f1e-5800-449c-b43d-01637f09083a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.730311] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 941.730547] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 941.730660] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Deleting the datastore file [datastore1] d2480bd6-527b-46b8-8a6a-22ad4eda3f5f {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.731056] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e72d4d6-21c8-4fce-8063-bb4ce42bba69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.739227] env[63355]: DEBUG oslo_vmware.api [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for the task: (returnval){ [ 941.739227] env[63355]: value = "task-1349807" [ 941.739227] env[63355]: _type = "Task" [ 941.739227] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.751693] env[63355]: DEBUG oslo_vmware.api [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349807, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.878187] env[63355]: DEBUG oslo_vmware.api [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349804, 'name': PowerOnVM_Task, 'duration_secs': 0.695485} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.878187] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.878535] env[63355]: INFO nova.compute.manager [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Took 10.88 seconds to spawn the instance on the hypervisor. [ 941.878535] env[63355]: DEBUG nova.compute.manager [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.884946] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d937c4ea-3c6c-4be0-b66f-35e6ace05675 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.940153] env[63355]: DEBUG nova.network.neutron [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Updating instance_info_cache with network_info: [{"id": "2f380960-ecf1-407f-aaf0-88703f8132c3", "address": "fa:16:3e:33:01:02", "network": {"id": "03000253-2cbc-4a28-aa72-31db532c2364", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1346884561", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.48", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad0fba7800a641da9114e93e18d7977d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f380960-ec", "ovs_interfaceid": "2f380960-ecf1-407f-aaf0-88703f8132c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cf363065-5623-4cf2-9d11-fcc3a234909a", "address": "fa:16:3e:fa:92:45", "network": {"id": "1032ad7b-9ec7-4d0e-bdd1-2f0f2e5d82f8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-908697110", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "ad0fba7800a641da9114e93e18d7977d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf363065-56", "ovs_interfaceid": "cf363065-5623-4cf2-9d11-fcc3a234909a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.055017] env[63355]: DEBUG nova.network.neutron [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Successfully updated port: 850ef0bb-ad5f-4627-b482-a78817b84343 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.061625] env[63355]: INFO nova.compute.manager [-] [instance: a408e19a-0960-430d-8550-0a304c63da61] Took 1.63 seconds to deallocate network for instance. [ 942.088529] env[63355]: DEBUG nova.compute.manager [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Received event network-vif-plugged-cf363065-5623-4cf2-9d11-fcc3a234909a {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.088936] env[63355]: DEBUG oslo_concurrency.lockutils [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] Acquiring lock "6706c65c-3959-471b-82c0-2196f4aa1ab0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.089934] env[63355]: DEBUG oslo_concurrency.lockutils [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] Lock "6706c65c-3959-471b-82c0-2196f4aa1ab0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.090221] env[63355]: DEBUG oslo_concurrency.lockutils [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] Lock "6706c65c-3959-471b-82c0-2196f4aa1ab0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.090503] env[63355]: DEBUG nova.compute.manager [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] No waiting events found dispatching network-vif-plugged-cf363065-5623-4cf2-9d11-fcc3a234909a {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 942.090732] env[63355]: WARNING nova.compute.manager [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Received unexpected event network-vif-plugged-cf363065-5623-4cf2-9d11-fcc3a234909a for instance with vm_state building and task_state spawning. [ 942.090989] env[63355]: DEBUG nova.compute.manager [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Received event network-changed-cf363065-5623-4cf2-9d11-fcc3a234909a {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.091348] env[63355]: DEBUG nova.compute.manager [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Refreshing instance network info cache due to event network-changed-cf363065-5623-4cf2-9d11-fcc3a234909a. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.091598] env[63355]: DEBUG oslo_concurrency.lockutils [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] Acquiring lock "refresh_cache-6706c65c-3959-471b-82c0-2196f4aa1ab0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.096289] env[63355]: DEBUG nova.compute.utils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.106310] env[63355]: DEBUG nova.compute.manager [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 942.106506] env[63355]: DEBUG nova.network.neutron [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 942.141989] env[63355]: WARNING nova.network.neutron [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] 943d3ecf-0c77-4c51-a997-b17d86259c34 already exists in list: networks containing: ['943d3ecf-0c77-4c51-a997-b17d86259c34']. ignoring it [ 942.141989] env[63355]: WARNING nova.network.neutron [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] 943d3ecf-0c77-4c51-a997-b17d86259c34 already exists in list: networks containing: ['943d3ecf-0c77-4c51-a997-b17d86259c34']. ignoring it [ 942.175821] env[63355]: DEBUG nova.policy [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bf70ebeeb3a4327ada51d18f5651476', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce0682a99ac94aeea463c961b84e6b58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 942.256792] env[63355]: DEBUG oslo_vmware.api [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Task: {'id': task-1349807, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.374177} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.257638] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.257638] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 942.257811] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 942.257930] env[63355]: INFO nova.compute.manager [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 942.258213] env[63355]: DEBUG oslo.service.loopingcall [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.258416] env[63355]: DEBUG nova.compute.manager [-] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 942.258512] env[63355]: DEBUG nova.network.neutron [-] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 942.412410] env[63355]: INFO nova.compute.manager [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Took 33.24 seconds to build instance. [ 942.445140] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Releasing lock "refresh_cache-6706c65c-3959-471b-82c0-2196f4aa1ab0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.445250] env[63355]: DEBUG nova.compute.manager [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Instance network_info: |[{"id": "2f380960-ecf1-407f-aaf0-88703f8132c3", "address": "fa:16:3e:33:01:02", "network": {"id": "03000253-2cbc-4a28-aa72-31db532c2364", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1346884561", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.48", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad0fba7800a641da9114e93e18d7977d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f380960-ec", "ovs_interfaceid": "2f380960-ecf1-407f-aaf0-88703f8132c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cf363065-5623-4cf2-9d11-fcc3a234909a", "address": "fa:16:3e:fa:92:45", "network": {"id": "1032ad7b-9ec7-4d0e-bdd1-2f0f2e5d82f8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-908697110", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "ad0fba7800a641da9114e93e18d7977d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf363065-56", "ovs_interfaceid": "cf363065-5623-4cf2-9d11-fcc3a234909a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.445537] env[63355]: DEBUG oslo_concurrency.lockutils [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] Acquired lock "refresh_cache-6706c65c-3959-471b-82c0-2196f4aa1ab0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.445730] env[63355]: DEBUG nova.network.neutron [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Refreshing network info cache for port cf363065-5623-4cf2-9d11-fcc3a234909a {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 942.447531] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:01:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c7d2575f-b92f-44ec-a863-634cb76631a2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2f380960-ecf1-407f-aaf0-88703f8132c3', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:92:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf363065-5623-4cf2-9d11-fcc3a234909a', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.461401] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Creating folder: Project (ad0fba7800a641da9114e93e18d7977d). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 942.464660] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-733abaeb-a789-4425-a212-4057315e74db {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.487560] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Created folder: Project (ad0fba7800a641da9114e93e18d7977d) in parent group-v287607. [ 942.487850] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Creating folder: Instances. Parent ref: group-v287698. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 942.490947] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-87c3c0f9-0341-4666-a67c-afa9eac339c1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.509451] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Created folder: Instances in parent group-v287698. [ 942.509796] env[63355]: DEBUG oslo.service.loopingcall [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.512570] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 942.513041] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63c24dda-d2ae-447f-8808-aa0cb7698745 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.534333] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04582396-6ca6-429e-884d-80ca3c9b4283 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.545274] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0b1e57-c9a5-47a3-8a7a-153b0cdcf091 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.549386] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.549386] env[63355]: value = "task-1349810" [ 942.549386] env[63355]: _type = "Task" [ 942.549386] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.578302] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.578461] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.578631] env[63355]: DEBUG nova.network.neutron [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.580466] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.584916] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf582df-49d4-4b4e-8836-212284ebbdc6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.587896] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349810, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.594782] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f6957b-8b47-4a2e-a107-5ba55f0ba04d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.610395] env[63355]: DEBUG nova.compute.manager [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 942.613493] env[63355]: DEBUG nova.compute.provider_tree [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.672138] env[63355]: DEBUG nova.network.neutron [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Successfully created port: 91c7c439-7d4c-4512-90c6-d1698630ae2b {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 942.916678] env[63355]: DEBUG oslo_concurrency.lockutils [None req-dfdf1764-8ece-4269-beb2-c33b920e491f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.478s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.063194] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349810, 'name': CreateVM_Task, 'duration_secs': 0.473703} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.063334] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 943.064230] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.064412] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.064747] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.065059] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee4daf05-ec71-4bc8-bf95-6638ed0e5e23 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.071498] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for the task: (returnval){ [ 943.071498] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521ad36b-227e-1f1c-54f5-2eabea295944" [ 943.071498] env[63355]: _type = "Task" [ 943.071498] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.081807] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521ad36b-227e-1f1c-54f5-2eabea295944, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.119709] env[63355]: DEBUG nova.scheduler.client.report [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.148166] env[63355]: DEBUG nova.network.neutron [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 943.251642] env[63355]: DEBUG nova.network.neutron [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Updated VIF entry in instance network info cache for port cf363065-5623-4cf2-9d11-fcc3a234909a. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 943.252094] env[63355]: DEBUG nova.network.neutron [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Updating instance_info_cache with network_info: [{"id": "2f380960-ecf1-407f-aaf0-88703f8132c3", "address": "fa:16:3e:33:01:02", "network": {"id": "03000253-2cbc-4a28-aa72-31db532c2364", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1346884561", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.48", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad0fba7800a641da9114e93e18d7977d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f380960-ec", "ovs_interfaceid": "2f380960-ecf1-407f-aaf0-88703f8132c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cf363065-5623-4cf2-9d11-fcc3a234909a", "address": "fa:16:3e:fa:92:45", "network": {"id": "1032ad7b-9ec7-4d0e-bdd1-2f0f2e5d82f8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-908697110", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "ad0fba7800a641da9114e93e18d7977d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf363065-56", "ovs_interfaceid": "cf363065-5623-4cf2-9d11-fcc3a234909a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.281450] env[63355]: DEBUG nova.network.neutron [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [{"id": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "address": "fa:16:3e:a9:13:18", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348dcf0-b9", "ovs_interfaceid": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "678307c8-eb2a-409d-a339-e40d5fd2ee50", "address": "fa:16:3e:9b:4c:95", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap678307c8-eb", "ovs_interfaceid": "678307c8-eb2a-409d-a339-e40d5fd2ee50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bd4c85c2-7841-4da0-ba95-43109bfd107e", "address": "fa:16:3e:27:d8:d6", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd4c85c2-78", "ovs_interfaceid": "bd4c85c2-7841-4da0-ba95-43109bfd107e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.390415] env[63355]: DEBUG nova.network.neutron [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Updating instance_info_cache with network_info: [{"id": "850ef0bb-ad5f-4627-b482-a78817b84343", "address": "fa:16:3e:42:ee:f7", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap850ef0bb-ad", "ovs_interfaceid": "850ef0bb-ad5f-4627-b482-a78817b84343", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.451878] env[63355]: DEBUG nova.network.neutron [-] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.586035] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521ad36b-227e-1f1c-54f5-2eabea295944, 'name': SearchDatastore_Task, 'duration_secs': 0.018364} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.586417] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.586731] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.588478] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.588759] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.588916] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.589138] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-999ac692-0470-4470-a63f-d2e4d8545338 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.603628] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.603840] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.604615] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cec0464c-87cf-4b77-8c0d-86a99861cbd7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.611519] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for the task: (returnval){ [ 943.611519] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52371621-40b3-6476-6912-64cfe299c63f" [ 943.611519] env[63355]: _type = "Task" [ 943.611519] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.627120] env[63355]: DEBUG nova.compute.manager [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 943.630074] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.038s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.631948] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52371621-40b3-6476-6912-64cfe299c63f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.634374] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.662s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.634374] env[63355]: DEBUG nova.objects.instance [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lazy-loading 'resources' on Instance uuid b4b09b1d-680e-47b8-aa8a-9b3d9167824d {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.666243] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 943.666531] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 943.666714] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.666940] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 943.671258] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.671567] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 943.671757] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 943.671848] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 943.672499] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 943.672499] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 943.672499] env[63355]: DEBUG nova.virt.hardware [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.673608] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1924824d-eb62-46d7-ab04-14f042c12f9a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.680467] env[63355]: INFO nova.scheduler.client.report [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Deleted allocations for instance 7c139710-d8d9-4cd7-bec0-6e021d3b2e68 [ 943.689291] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48700bb8-4107-4c6e-96bb-37f8fd9a1828 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.754944] env[63355]: DEBUG oslo_concurrency.lockutils [req-ccb552dd-e404-4681-8112-976fc756963e req-223f5637-5855-4c4f-a381-7a011071e424 service nova] Releasing lock "refresh_cache-6706c65c-3959-471b-82c0-2196f4aa1ab0" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.785609] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.786430] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.786610] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.787791] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de78b76-a05a-4fa1-82bb-575c1432ec3d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.816375] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 943.817842] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 943.817842] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.817842] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 943.819724] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.819724] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 943.819724] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 943.819724] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 943.819724] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 943.819724] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 943.819724] env[63355]: DEBUG nova.virt.hardware [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.826917] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Reconfiguring VM to attach interface {{(pid=63355) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 943.827387] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efb87b92-f895-461e-83be-753516643d91 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.850172] env[63355]: DEBUG oslo_vmware.api [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 943.850172] env[63355]: value = "task-1349811" [ 943.850172] env[63355]: _type = "Task" [ 943.850172] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.862853] env[63355]: DEBUG oslo_vmware.api [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349811, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.895201] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.895626] env[63355]: DEBUG nova.compute.manager [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Instance network_info: |[{"id": "850ef0bb-ad5f-4627-b482-a78817b84343", "address": "fa:16:3e:42:ee:f7", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap850ef0bb-ad", "ovs_interfaceid": "850ef0bb-ad5f-4627-b482-a78817b84343", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 943.896183] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:ee:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '52358fcc-0d9f-45dd-8c75-db533fd992c3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '850ef0bb-ad5f-4627-b482-a78817b84343', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.904408] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Creating folder: Project (cfff1b9903264e5586119ebd3a3602de). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 943.904743] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7364d53a-e328-44c7-965f-7f989e2a6ef6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.918919] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Created folder: Project (cfff1b9903264e5586119ebd3a3602de) in parent group-v287607. [ 943.919262] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Creating folder: Instances. Parent ref: group-v287701. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 943.919609] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ebabdda3-45b4-475d-b06e-49a75948110c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.933546] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Created folder: Instances in parent group-v287701. [ 943.934095] env[63355]: DEBUG oslo.service.loopingcall [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.934095] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.934328] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ab5f853-f4c9-4787-9896-ef1c0d649706 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.954852] env[63355]: INFO nova.compute.manager [-] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Took 1.70 seconds to deallocate network for instance. [ 943.958239] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.958239] env[63355]: value = "task-1349814" [ 943.958239] env[63355]: _type = "Task" [ 943.958239] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.973052] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349814, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.124864] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52371621-40b3-6476-6912-64cfe299c63f, 'name': SearchDatastore_Task, 'duration_secs': 0.015806} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.125849] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7aef0cd7-3145-4b70-95b8-ca7550bed94b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.132694] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for the task: (returnval){ [ 944.132694] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5267b4a1-5a2e-fd17-b1bf-d2b482d2caf4" [ 944.132694] env[63355]: _type = "Task" [ 944.132694] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.146238] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5267b4a1-5a2e-fd17-b1bf-d2b482d2caf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.203647] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9f820235-d7d4-4fac-8095-908530ac2491 tempest-InstanceActionsNegativeTestJSON-93827226 tempest-InstanceActionsNegativeTestJSON-93827226-project-member] Lock "7c139710-d8d9-4cd7-bec0-6e021d3b2e68" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.921s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.327121] env[63355]: DEBUG nova.compute.manager [req-ddee96d9-8904-4ad7-96df-b8760a45be88 req-769e718d-df66-4ee5-a4ee-0f4182bd6d8d service nova] [instance: a408e19a-0960-430d-8550-0a304c63da61] Received event network-vif-deleted-de9150b0-5d81-4397-a0d7-58c9a471d791 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.327494] env[63355]: DEBUG nova.compute.manager [req-ddee96d9-8904-4ad7-96df-b8760a45be88 req-769e718d-df66-4ee5-a4ee-0f4182bd6d8d service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received event network-changed-bd4c85c2-7841-4da0-ba95-43109bfd107e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.327827] env[63355]: DEBUG nova.compute.manager [req-ddee96d9-8904-4ad7-96df-b8760a45be88 req-769e718d-df66-4ee5-a4ee-0f4182bd6d8d service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Refreshing instance network info cache due to event network-changed-bd4c85c2-7841-4da0-ba95-43109bfd107e. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 944.328182] env[63355]: DEBUG oslo_concurrency.lockutils [req-ddee96d9-8904-4ad7-96df-b8760a45be88 req-769e718d-df66-4ee5-a4ee-0f4182bd6d8d service nova] Acquiring lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.328454] env[63355]: DEBUG oslo_concurrency.lockutils [req-ddee96d9-8904-4ad7-96df-b8760a45be88 req-769e718d-df66-4ee5-a4ee-0f4182bd6d8d service nova] Acquired lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.328725] env[63355]: DEBUG nova.network.neutron [req-ddee96d9-8904-4ad7-96df-b8760a45be88 req-769e718d-df66-4ee5-a4ee-0f4182bd6d8d service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Refreshing network info cache for port bd4c85c2-7841-4da0-ba95-43109bfd107e {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 944.366031] env[63355]: DEBUG oslo_vmware.api [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349811, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.471549] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.479647] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349814, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.486314] env[63355]: DEBUG nova.compute.manager [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Received event network-vif-plugged-850ef0bb-ad5f-4627-b482-a78817b84343 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.486568] env[63355]: DEBUG oslo_concurrency.lockutils [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] Acquiring lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.486733] env[63355]: DEBUG oslo_concurrency.lockutils [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.486940] env[63355]: DEBUG oslo_concurrency.lockutils [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.487262] env[63355]: DEBUG nova.compute.manager [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] No waiting events found dispatching network-vif-plugged-850ef0bb-ad5f-4627-b482-a78817b84343 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 944.487470] env[63355]: WARNING nova.compute.manager [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Received unexpected event network-vif-plugged-850ef0bb-ad5f-4627-b482-a78817b84343 for instance with vm_state building and task_state spawning. [ 944.487634] env[63355]: DEBUG nova.compute.manager [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Received event network-changed-850ef0bb-ad5f-4627-b482-a78817b84343 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.487786] env[63355]: DEBUG nova.compute.manager [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Refreshing instance network info cache due to event network-changed-850ef0bb-ad5f-4627-b482-a78817b84343. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 944.488010] env[63355]: DEBUG oslo_concurrency.lockutils [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] Acquiring lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.488197] env[63355]: DEBUG oslo_concurrency.lockutils [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] Acquired lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.488368] env[63355]: DEBUG nova.network.neutron [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Refreshing network info cache for port 850ef0bb-ad5f-4627-b482-a78817b84343 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 944.495402] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363e667b-48bd-4873-b903-99bd47ec216e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.505830] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b402c0-e7fc-4f57-a11f-d8d5a50bc2e5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.540791] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca36913f-d2a1-4c92-9e36-298f8bda5ce6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.549613] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16595e40-8a0a-47d3-a800-19b3c6defb4b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.567297] env[63355]: DEBUG nova.compute.provider_tree [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.641762] env[63355]: DEBUG nova.network.neutron [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Successfully updated port: 91c7c439-7d4c-4512-90c6-d1698630ae2b {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 944.647435] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5267b4a1-5a2e-fd17-b1bf-d2b482d2caf4, 'name': SearchDatastore_Task, 'duration_secs': 0.012094} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.649646] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.649646] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 6706c65c-3959-471b-82c0-2196f4aa1ab0/6706c65c-3959-471b-82c0-2196f4aa1ab0.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.649646] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f22327fd-d9d0-4763-9e8c-ee6c24f19381 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.658659] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for the task: (returnval){ [ 944.658659] env[63355]: value = "task-1349815" [ 944.658659] env[63355]: _type = "Task" [ 944.658659] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.670912] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349815, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.865770] env[63355]: DEBUG oslo_vmware.api [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349811, 'name': ReconfigVM_Task, 'duration_secs': 0.817264} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.865770] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.865770] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Reconfigured VM to attach interface {{(pid=63355) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 944.982576] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349814, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.071255] env[63355]: DEBUG nova.scheduler.client.report [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.148845] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-1570cb36-76e0-4d06-8080-735b5246e92e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.148940] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-1570cb36-76e0-4d06-8080-735b5246e92e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.150351] env[63355]: DEBUG nova.network.neutron [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 945.182395] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349815, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521412} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.182895] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 6706c65c-3959-471b-82c0-2196f4aa1ab0/6706c65c-3959-471b-82c0-2196f4aa1ab0.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.183112] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.183570] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07e544ed-a6fe-40c6-861f-7ba77abf2788 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.192315] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for the task: (returnval){ [ 945.192315] env[63355]: value = "task-1349816" [ 945.192315] env[63355]: _type = "Task" [ 945.192315] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.204103] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349816, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.208446] env[63355]: DEBUG nova.network.neutron [req-ddee96d9-8904-4ad7-96df-b8760a45be88 req-769e718d-df66-4ee5-a4ee-0f4182bd6d8d service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updated VIF entry in instance network info cache for port bd4c85c2-7841-4da0-ba95-43109bfd107e. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.209073] env[63355]: DEBUG nova.network.neutron [req-ddee96d9-8904-4ad7-96df-b8760a45be88 req-769e718d-df66-4ee5-a4ee-0f4182bd6d8d service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [{"id": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "address": "fa:16:3e:a9:13:18", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348dcf0-b9", "ovs_interfaceid": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "678307c8-eb2a-409d-a339-e40d5fd2ee50", "address": "fa:16:3e:9b:4c:95", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap678307c8-eb", "ovs_interfaceid": "678307c8-eb2a-409d-a339-e40d5fd2ee50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bd4c85c2-7841-4da0-ba95-43109bfd107e", "address": "fa:16:3e:27:d8:d6", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd4c85c2-78", "ovs_interfaceid": "bd4c85c2-7841-4da0-ba95-43109bfd107e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.318074] env[63355]: DEBUG nova.network.neutron [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Updated VIF entry in instance network info cache for port 850ef0bb-ad5f-4627-b482-a78817b84343. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.318474] env[63355]: DEBUG nova.network.neutron [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Updating instance_info_cache with network_info: [{"id": "850ef0bb-ad5f-4627-b482-a78817b84343", "address": "fa:16:3e:42:ee:f7", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap850ef0bb-ad", "ovs_interfaceid": "850ef0bb-ad5f-4627-b482-a78817b84343", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.368763] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9de34bfd-dc62-4f0e-ae74-0521f9af00c7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-386e847e-967b-4247-9730-cdc5ac251474-bd4c85c2-7841-4da0-ba95-43109bfd107e" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.664s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.477289] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349814, 'name': CreateVM_Task, 'duration_secs': 1.383448} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.477591] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.478392] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.479151] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.479151] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 945.479276] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92fc86ba-efbe-4420-b2dd-99bd9952b3bd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.485753] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 945.485753] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d1240b-111a-b8af-5506-a1c568ae5501" [ 945.485753] env[63355]: _type = "Task" [ 945.485753] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.496739] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d1240b-111a-b8af-5506-a1c568ae5501, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.529718] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.529992] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.576688] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.943s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.579284] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.430s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.581091] env[63355]: INFO nova.compute.claims [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.609615] env[63355]: INFO nova.scheduler.client.report [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Deleted allocations for instance b4b09b1d-680e-47b8-aa8a-9b3d9167824d [ 945.697155] env[63355]: DEBUG nova.network.neutron [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 945.708988] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349816, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072601} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.709306] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.713024] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb8513e-529e-4d73-813f-49a8757c47a4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.713176] env[63355]: DEBUG oslo_concurrency.lockutils [req-ddee96d9-8904-4ad7-96df-b8760a45be88 req-769e718d-df66-4ee5-a4ee-0f4182bd6d8d service nova] Releasing lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.713346] env[63355]: DEBUG nova.compute.manager [req-ddee96d9-8904-4ad7-96df-b8760a45be88 req-769e718d-df66-4ee5-a4ee-0f4182bd6d8d service nova] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Received event network-vif-deleted-8c82e15c-a3e6-4ec2-b00e-cc57c343b635 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.738909] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 6706c65c-3959-471b-82c0-2196f4aa1ab0/6706c65c-3959-471b-82c0-2196f4aa1ab0.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.743963] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-257a6f66-32ad-4523-afc2-8897c33b2a38 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.767894] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for the task: (returnval){ [ 945.767894] env[63355]: value = "task-1349817" [ 945.767894] env[63355]: _type = "Task" [ 945.767894] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.776638] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349817, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.821890] env[63355]: DEBUG oslo_concurrency.lockutils [req-547cc6ec-6a5a-440c-a152-6085bf59fb8e req-ddb153d5-f8de-4fc2-8d94-6d8194d35bd8 service nova] Releasing lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.890482] env[63355]: DEBUG nova.network.neutron [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Updating instance_info_cache with network_info: [{"id": "91c7c439-7d4c-4512-90c6-d1698630ae2b", "address": "fa:16:3e:14:1e:d7", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91c7c439-7d", "ovs_interfaceid": "91c7c439-7d4c-4512-90c6-d1698630ae2b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.001758] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d1240b-111a-b8af-5506-a1c568ae5501, 'name': SearchDatastore_Task, 'duration_secs': 0.0108} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.002263] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.002588] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.002763] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.002964] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.003453] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.004801] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88b5a637-4d54-49fa-8062-2ede8560f102 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.018525] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.018895] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.020032] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf789f71-029e-47ce-9d76-dedf24a86ea8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.027704] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 946.027704] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52dd4e44-599b-b751-fe3d-e590a10da81d" [ 946.027704] env[63355]: _type = "Task" [ 946.027704] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.033985] env[63355]: DEBUG nova.compute.manager [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 946.047610] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52dd4e44-599b-b751-fe3d-e590a10da81d, 'name': SearchDatastore_Task, 'duration_secs': 0.013217} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.049077] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e2b43fd-80d1-4abb-9c97-b86fbc5e130a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.059184] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 946.059184] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52348c50-db9a-fdc7-a7ec-24192f72b6c1" [ 946.059184] env[63355]: _type = "Task" [ 946.059184] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.074341] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52348c50-db9a-fdc7-a7ec-24192f72b6c1, 'name': SearchDatastore_Task, 'duration_secs': 0.012764} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.074682] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.075137] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] cd9de9b8-ad42-4dbb-b435-927738a55f3f/cd9de9b8-ad42-4dbb-b435-927738a55f3f.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 946.075345] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8aa0d73-4d16-4e97-8a02-7272ae789618 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.088972] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 946.088972] env[63355]: value = "task-1349818" [ 946.088972] env[63355]: _type = "Task" [ 946.088972] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.100572] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349818, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.122101] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e551488d-03a2-4c42-961d-b94943c69020 tempest-ServerShowV257Test-1397437046 tempest-ServerShowV257Test-1397437046-project-member] Lock "b4b09b1d-680e-47b8-aa8a-9b3d9167824d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.986s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.283329] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349817, 'name': ReconfigVM_Task, 'duration_secs': 0.308978} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.284010] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 6706c65c-3959-471b-82c0-2196f4aa1ab0/6706c65c-3959-471b-82c0-2196f4aa1ab0.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.284809] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13958690-d075-409e-8c13-694e9a6c9d72 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.294825] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for the task: (returnval){ [ 946.294825] env[63355]: value = "task-1349819" [ 946.294825] env[63355]: _type = "Task" [ 946.294825] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.304824] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349819, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.372574] env[63355]: DEBUG nova.compute.manager [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Received event network-vif-plugged-91c7c439-7d4c-4512-90c6-d1698630ae2b {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.372866] env[63355]: DEBUG oslo_concurrency.lockutils [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] Acquiring lock "1570cb36-76e0-4d06-8080-735b5246e92e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.373251] env[63355]: DEBUG oslo_concurrency.lockutils [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] Lock "1570cb36-76e0-4d06-8080-735b5246e92e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.373574] env[63355]: DEBUG oslo_concurrency.lockutils [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] Lock "1570cb36-76e0-4d06-8080-735b5246e92e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.373849] env[63355]: DEBUG nova.compute.manager [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] No waiting events found dispatching network-vif-plugged-91c7c439-7d4c-4512-90c6-d1698630ae2b {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 946.374186] env[63355]: WARNING nova.compute.manager [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Received unexpected event network-vif-plugged-91c7c439-7d4c-4512-90c6-d1698630ae2b for instance with vm_state building and task_state spawning. [ 946.374483] env[63355]: DEBUG nova.compute.manager [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Received event network-changed-91c7c439-7d4c-4512-90c6-d1698630ae2b {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.374883] env[63355]: DEBUG nova.compute.manager [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Refreshing instance network info cache due to event network-changed-91c7c439-7d4c-4512-90c6-d1698630ae2b. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.375397] env[63355]: DEBUG oslo_concurrency.lockutils [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] Acquiring lock "refresh_cache-1570cb36-76e0-4d06-8080-735b5246e92e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.392131] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-1570cb36-76e0-4d06-8080-735b5246e92e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.392485] env[63355]: DEBUG nova.compute.manager [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Instance network_info: |[{"id": "91c7c439-7d4c-4512-90c6-d1698630ae2b", "address": "fa:16:3e:14:1e:d7", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91c7c439-7d", "ovs_interfaceid": "91c7c439-7d4c-4512-90c6-d1698630ae2b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 946.392886] env[63355]: DEBUG oslo_concurrency.lockutils [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] Acquired lock "refresh_cache-1570cb36-76e0-4d06-8080-735b5246e92e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.393125] env[63355]: DEBUG nova.network.neutron [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Refreshing network info cache for port 91c7c439-7d4c-4512-90c6-d1698630ae2b {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.394578] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:1e:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '130387c4-e4ec-4d95-8e9d-bb079baabad8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91c7c439-7d4c-4512-90c6-d1698630ae2b', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.403727] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Creating folder: Project (ce0682a99ac94aeea463c961b84e6b58). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.403727] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da393a86-bef4-4041-bc96-e462c51460e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.419409] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Created folder: Project (ce0682a99ac94aeea463c961b84e6b58) in parent group-v287607. [ 946.420021] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Creating folder: Instances. Parent ref: group-v287704. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.420218] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2406bec4-6bd5-4c52-a37b-a047b833f197 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.438068] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Created folder: Instances in parent group-v287704. [ 946.438068] env[63355]: DEBUG oslo.service.loopingcall [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.438068] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.438068] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a32ef412-28d4-4714-941c-da648a1be285 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.458572] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.458572] env[63355]: value = "task-1349822" [ 946.458572] env[63355]: _type = "Task" [ 946.458572] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.470152] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349822, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.569135] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.607880] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349818, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.705560] env[63355]: DEBUG nova.compute.manager [req-82266d88-0a6a-4d52-8b52-0451f3e0dae0 req-2abe3519-ff52-4280-b2eb-ccc8df33cbae service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received event network-changed-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.705560] env[63355]: DEBUG nova.compute.manager [req-82266d88-0a6a-4d52-8b52-0451f3e0dae0 req-2abe3519-ff52-4280-b2eb-ccc8df33cbae service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Refreshing instance network info cache due to event network-changed-43ebe841-d4c8-446a-981c-519bef977228. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.705560] env[63355]: DEBUG oslo_concurrency.lockutils [req-82266d88-0a6a-4d52-8b52-0451f3e0dae0 req-2abe3519-ff52-4280-b2eb-ccc8df33cbae service nova] Acquiring lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.705560] env[63355]: DEBUG oslo_concurrency.lockutils [req-82266d88-0a6a-4d52-8b52-0451f3e0dae0 req-2abe3519-ff52-4280-b2eb-ccc8df33cbae service nova] Acquired lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.705560] env[63355]: DEBUG nova.network.neutron [req-82266d88-0a6a-4d52-8b52-0451f3e0dae0 req-2abe3519-ff52-4280-b2eb-ccc8df33cbae service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Refreshing network info cache for port 43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.809688] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349819, 'name': Rename_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.945750] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c633fe-1354-4a0d-bc21-65459356f1fc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.954994] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2b4352-57ae-4b34-b6f3-f6e51f0ba780 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.994734] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519c6c78-b213-4271-b201-b33f1359c4f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.004075] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349822, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.016176] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ac4d89-fe09-48ea-9f36-3f18bbb7a308 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.033184] env[63355]: DEBUG nova.compute.provider_tree [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.103554] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349818, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602632} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.103724] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] cd9de9b8-ad42-4dbb-b435-927738a55f3f/cd9de9b8-ad42-4dbb-b435-927738a55f3f.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 947.103953] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.104251] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3695f169-4620-4876-b049-6a61262b7ddc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.112504] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 947.112504] env[63355]: value = "task-1349823" [ 947.112504] env[63355]: _type = "Task" [ 947.112504] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.128061] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349823, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.197912] env[63355]: DEBUG nova.network.neutron [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Updated VIF entry in instance network info cache for port 91c7c439-7d4c-4512-90c6-d1698630ae2b. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.198383] env[63355]: DEBUG nova.network.neutron [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Updating instance_info_cache with network_info: [{"id": "91c7c439-7d4c-4512-90c6-d1698630ae2b", "address": "fa:16:3e:14:1e:d7", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91c7c439-7d", "ovs_interfaceid": "91c7c439-7d4c-4512-90c6-d1698630ae2b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.306761] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349819, 'name': Rename_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.456484] env[63355]: DEBUG nova.network.neutron [req-82266d88-0a6a-4d52-8b52-0451f3e0dae0 req-2abe3519-ff52-4280-b2eb-ccc8df33cbae service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updated VIF entry in instance network info cache for port 43ebe841-d4c8-446a-981c-519bef977228. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.456484] env[63355]: DEBUG nova.network.neutron [req-82266d88-0a6a-4d52-8b52-0451f3e0dae0 req-2abe3519-ff52-4280-b2eb-ccc8df33cbae service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updating instance_info_cache with network_info: [{"id": "43ebe841-d4c8-446a-981c-519bef977228", "address": "fa:16:3e:97:62:7e", "network": {"id": "e19b4619-6e81-4205-838d-9ee7e7ae44d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1968447346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "744ce7dc0ab4463fa461cbac87e51904", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43ebe841-d4", "ovs_interfaceid": "43ebe841-d4c8-446a-981c-519bef977228", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.471985] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349822, 'name': CreateVM_Task, 'duration_secs': 0.606155} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.472195] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.472922] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.473136] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.473479] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.474114] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-843a1bb0-bb28-49b7-a911-52b3ad63862c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.480146] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 947.480146] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a1e897-a8d3-1527-84fb-de7bb00c979b" [ 947.480146] env[63355]: _type = "Task" [ 947.480146] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.489628] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a1e897-a8d3-1527-84fb-de7bb00c979b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.503663] env[63355]: DEBUG oslo_concurrency.lockutils [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "interface-386e847e-967b-4247-9730-cdc5ac251474-678307c8-eb2a-409d-a339-e40d5fd2ee50" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.503878] env[63355]: DEBUG oslo_concurrency.lockutils [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-386e847e-967b-4247-9730-cdc5ac251474-678307c8-eb2a-409d-a339-e40d5fd2ee50" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.538945] env[63355]: DEBUG nova.scheduler.client.report [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.625298] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349823, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073295} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.625656] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 947.626491] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1add2be6-3486-4d6e-bd2f-5f0116977e0d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.649354] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] cd9de9b8-ad42-4dbb-b435-927738a55f3f/cd9de9b8-ad42-4dbb-b435-927738a55f3f.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.649665] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee48465e-a000-492f-8287-acf64745f070 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.671116] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 947.671116] env[63355]: value = "task-1349824" [ 947.671116] env[63355]: _type = "Task" [ 947.671116] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.680748] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349824, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.701187] env[63355]: DEBUG oslo_concurrency.lockutils [req-da3447be-29e0-4a6d-9143-c03eb2980d52 req-b8a99791-7189-4bc0-ab24-cfd927af2064 service nova] Releasing lock "refresh_cache-1570cb36-76e0-4d06-8080-735b5246e92e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.807404] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349819, 'name': Rename_Task, 'duration_secs': 1.301939} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.808242] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.808242] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b14c421-9496-4ebb-bdb6-df8ea29ef12d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.815996] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for the task: (returnval){ [ 947.815996] env[63355]: value = "task-1349825" [ 947.815996] env[63355]: _type = "Task" [ 947.815996] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.825085] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349825, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.958152] env[63355]: DEBUG oslo_concurrency.lockutils [req-82266d88-0a6a-4d52-8b52-0451f3e0dae0 req-2abe3519-ff52-4280-b2eb-ccc8df33cbae service nova] Releasing lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.987284] env[63355]: DEBUG oslo_vmware.rw_handles [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ba933-dea8-3806-76b5-deed25ade1ba/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 947.987829] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aca9080-8ee6-4f4e-bd49-2f876afc9420 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.997041] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a1e897-a8d3-1527-84fb-de7bb00c979b, 'name': SearchDatastore_Task, 'duration_secs': 0.017614} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.998782] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.999042] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.999290] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.999435] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.999611] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.999903] env[63355]: DEBUG oslo_vmware.rw_handles [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ba933-dea8-3806-76b5-deed25ade1ba/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 948.000062] env[63355]: ERROR oslo_vmware.rw_handles [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ba933-dea8-3806-76b5-deed25ade1ba/disk-0.vmdk due to incomplete transfer. [ 948.000298] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c86db5f-34e2-489d-928a-b23b9e70a5d1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.002168] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-44a7d7f8-3a91-46bc-9a9b-bb7290207737 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.006621] env[63355]: DEBUG oslo_concurrency.lockutils [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.006836] env[63355]: DEBUG oslo_concurrency.lockutils [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.007632] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55e6c0c-6b7d-4d5f-87ad-ec1fe9e50794 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.014061] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.014317] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 948.015049] env[63355]: DEBUG oslo_vmware.rw_handles [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ba933-dea8-3806-76b5-deed25ade1ba/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 948.015301] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Uploaded image faa91285-130f-41be-b93d-fe59ddde3549 to the Glance image server {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 948.017780] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 948.018568] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f716d30-327a-41a3-a834-30311b45b975 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.021152] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-895bfd51-d9e6-4b3c-a1b8-2690715569b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.037889] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7578050-0c45-42c1-ae47-6edd1a61db07 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.042912] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 948.042912] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528a604e-3431-cc2f-a969-c80d4d2df7f2" [ 948.042912] env[63355]: _type = "Task" [ 948.042912] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.045189] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.045445] env[63355]: DEBUG nova.compute.manager [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 948.049092] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 948.049092] env[63355]: value = "task-1349826" [ 948.049092] env[63355]: _type = "Task" [ 948.049092] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.071924] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.423s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.073498] env[63355]: INFO nova.compute.claims [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 948.081656] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Reconfiguring VM to detach interface {{(pid=63355) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 948.086031] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f4e683b-7edd-4c2f-8b84-2ddbb877ad61 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.111424] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349826, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.112616] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528a604e-3431-cc2f-a969-c80d4d2df7f2, 'name': SearchDatastore_Task, 'duration_secs': 0.013831} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.113699] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 948.113699] env[63355]: value = "task-1349827" [ 948.113699] env[63355]: _type = "Task" [ 948.113699] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.113971] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1df6e34a-9918-48f7-8c76-18c772e23946 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.123622] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 948.123622] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529e992e-38a0-6fd0-ee69-cf7774be17d2" [ 948.123622] env[63355]: _type = "Task" [ 948.123622] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.127542] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.136761] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529e992e-38a0-6fd0-ee69-cf7774be17d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.184670] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349824, 'name': ReconfigVM_Task, 'duration_secs': 0.367109} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.184670] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Reconfigured VM instance instance-0000004e to attach disk [datastore2] cd9de9b8-ad42-4dbb-b435-927738a55f3f/cd9de9b8-ad42-4dbb-b435-927738a55f3f.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.184670] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f7d00e9-c97d-479c-8737-30b73f72535b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.193772] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 948.193772] env[63355]: value = "task-1349828" [ 948.193772] env[63355]: _type = "Task" [ 948.193772] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.204493] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349828, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.327340] env[63355]: DEBUG oslo_vmware.api [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349825, 'name': PowerOnVM_Task, 'duration_secs': 0.508014} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.327589] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.327807] env[63355]: INFO nova.compute.manager [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Took 12.76 seconds to spawn the instance on the hypervisor. [ 948.328034] env[63355]: DEBUG nova.compute.manager [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.328871] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5710949a-d5b7-4e16-bae7-d7d2c32cd608 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.570152] env[63355]: DEBUG nova.compute.utils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.571590] env[63355]: DEBUG nova.compute.manager [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 948.571756] env[63355]: DEBUG nova.network.neutron [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 948.588486] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349826, 'name': Destroy_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.627901] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.629524] env[63355]: DEBUG nova.policy [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3721c4a73df54ae9b44110cb9e8590a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc547065748241e8ac7b6c499ddaea66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.639726] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529e992e-38a0-6fd0-ee69-cf7774be17d2, 'name': SearchDatastore_Task, 'duration_secs': 0.017489} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.639968] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.640254] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 1570cb36-76e0-4d06-8080-735b5246e92e/1570cb36-76e0-4d06-8080-735b5246e92e.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 948.640515] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40b07f2f-048f-419a-9fc5-7cec2d647574 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.647587] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 948.647587] env[63355]: value = "task-1349829" [ 948.647587] env[63355]: _type = "Task" [ 948.647587] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.655992] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349829, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.704646] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349828, 'name': Rename_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.853352] env[63355]: INFO nova.compute.manager [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Took 33.89 seconds to build instance. [ 948.892583] env[63355]: DEBUG nova.network.neutron [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Successfully created port: f2f3c304-8432-48e3-a3ce-70547c912546 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 949.077063] env[63355]: DEBUG nova.compute.manager [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 949.092210] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349826, 'name': Destroy_Task, 'duration_secs': 0.571926} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.092684] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Destroyed the VM [ 949.092942] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 949.093233] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c3efbc72-3a29-4456-8fd4-95babbd6fe9f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.102401] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 949.102401] env[63355]: value = "task-1349830" [ 949.102401] env[63355]: _type = "Task" [ 949.102401] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.112774] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349830, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.128750] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.161222] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349829, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.211142] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349828, 'name': Rename_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.355675] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e1bc00b-90a7-4737-a20e-923ba52fa557 tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "6706c65c-3959-471b-82c0-2196f4aa1ab0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.915s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.396857] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8031eb94-87cf-4020-a560-4184f28098ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.408553] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129071c0-741b-456d-809b-e8ce8fa01418 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.454156] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25ce97d-ac63-4002-9f49-0a86c7274633 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.465407] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8461694-93bf-48ca-afd1-1449a91873cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.487976] env[63355]: DEBUG nova.compute.provider_tree [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.613442] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349830, 'name': RemoveSnapshot_Task} progress is 50%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.630123] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.647442] env[63355]: DEBUG oslo_concurrency.lockutils [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "6706c65c-3959-471b-82c0-2196f4aa1ab0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.647700] env[63355]: DEBUG oslo_concurrency.lockutils [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "6706c65c-3959-471b-82c0-2196f4aa1ab0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.647914] env[63355]: DEBUG oslo_concurrency.lockutils [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "6706c65c-3959-471b-82c0-2196f4aa1ab0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.648239] env[63355]: DEBUG oslo_concurrency.lockutils [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "6706c65c-3959-471b-82c0-2196f4aa1ab0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.648534] env[63355]: DEBUG oslo_concurrency.lockutils [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "6706c65c-3959-471b-82c0-2196f4aa1ab0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.653765] env[63355]: INFO nova.compute.manager [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Terminating instance [ 949.656072] env[63355]: DEBUG nova.compute.manager [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.656325] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.657102] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa89655b-5a4f-40b6-bba2-21a56a91cfdd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.663289] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349829, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.875772} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.664221] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 1570cb36-76e0-4d06-8080-735b5246e92e/1570cb36-76e0-4d06-8080-735b5246e92e.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 949.664471] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.664732] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-716c3089-59c8-4ae4-833b-42361ab35f23 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.670895] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.671461] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-222bdd26-da13-471a-b393-ac517a377734 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.676855] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 949.676855] env[63355]: value = "task-1349831" [ 949.676855] env[63355]: _type = "Task" [ 949.676855] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.680688] env[63355]: DEBUG oslo_vmware.api [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for the task: (returnval){ [ 949.680688] env[63355]: value = "task-1349832" [ 949.680688] env[63355]: _type = "Task" [ 949.680688] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.686946] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349831, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.694078] env[63355]: DEBUG oslo_vmware.api [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349832, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.704041] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349828, 'name': Rename_Task, 'duration_secs': 1.23753} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.704635] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 949.704709] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a450144-e5cc-4f08-8a28-14b0f27a4ebe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.712333] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 949.712333] env[63355]: value = "task-1349833" [ 949.712333] env[63355]: _type = "Task" [ 949.712333] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.722441] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349833, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.991364] env[63355]: DEBUG nova.scheduler.client.report [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.091643] env[63355]: DEBUG nova.compute.manager [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 950.113799] env[63355]: DEBUG oslo_vmware.api [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349830, 'name': RemoveSnapshot_Task, 'duration_secs': 0.633013} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.115959] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 950.116244] env[63355]: INFO nova.compute.manager [None req-61bcacaf-c3d8-47f9-96e0-86a8e44ad5ee tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Took 16.63 seconds to snapshot the instance on the hypervisor. [ 950.126505] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.126707] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.126872] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.127070] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.127225] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.127371] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.127584] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.127730] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.127905] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.128094] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.128305] env[63355]: DEBUG nova.virt.hardware [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.129656] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db134de2-0485-4385-b535-c90071834ee3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.134762] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.140544] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0657bd6-2818-4d2f-a8f5-3a3d6b2b2ca7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.190099] env[63355]: DEBUG oslo_vmware.api [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349832, 'name': PowerOffVM_Task, 'duration_secs': 0.253286} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.193092] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 950.193281] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 950.193559] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349831, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091489} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.194198] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3aa91d5-a69f-4634-ab83-7d2a5c71e450 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.195323] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.196121] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698bfc09-f1a7-44c6-94dc-ab8d8d6695ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.220477] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 1570cb36-76e0-4d06-8080-735b5246e92e/1570cb36-76e0-4d06-8080-735b5246e92e.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.220847] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8611c055-69b3-4676-b01e-b83c9f9a6f43 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.253115] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349833, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.254993] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 950.254993] env[63355]: value = "task-1349835" [ 950.254993] env[63355]: _type = "Task" [ 950.254993] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.266292] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349835, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.325635] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 950.325946] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 950.326180] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Deleting the datastore file [datastore2] 6706c65c-3959-471b-82c0-2196f4aa1ab0 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.326468] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-522209cb-1cf3-4db4-bf37-60a438e3ac4c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.335445] env[63355]: DEBUG oslo_vmware.api [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for the task: (returnval){ [ 950.335445] env[63355]: value = "task-1349836" [ 950.335445] env[63355]: _type = "Task" [ 950.335445] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.347272] env[63355]: DEBUG oslo_vmware.api [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349836, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.364653] env[63355]: DEBUG nova.compute.manager [req-de13830c-7a02-487b-90b4-8ed61435b85a req-4b2be61c-c49f-4260-9b63-e12db030f2ee service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Received event network-vif-plugged-f2f3c304-8432-48e3-a3ce-70547c912546 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.364843] env[63355]: DEBUG oslo_concurrency.lockutils [req-de13830c-7a02-487b-90b4-8ed61435b85a req-4b2be61c-c49f-4260-9b63-e12db030f2ee service nova] Acquiring lock "fc6a99e6-5319-47d3-8175-770c06d3e325-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.365112] env[63355]: DEBUG oslo_concurrency.lockutils [req-de13830c-7a02-487b-90b4-8ed61435b85a req-4b2be61c-c49f-4260-9b63-e12db030f2ee service nova] Lock "fc6a99e6-5319-47d3-8175-770c06d3e325-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.365330] env[63355]: DEBUG oslo_concurrency.lockutils [req-de13830c-7a02-487b-90b4-8ed61435b85a req-4b2be61c-c49f-4260-9b63-e12db030f2ee service nova] Lock "fc6a99e6-5319-47d3-8175-770c06d3e325-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.365532] env[63355]: DEBUG nova.compute.manager [req-de13830c-7a02-487b-90b4-8ed61435b85a req-4b2be61c-c49f-4260-9b63-e12db030f2ee service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] No waiting events found dispatching network-vif-plugged-f2f3c304-8432-48e3-a3ce-70547c912546 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 950.365747] env[63355]: WARNING nova.compute.manager [req-de13830c-7a02-487b-90b4-8ed61435b85a req-4b2be61c-c49f-4260-9b63-e12db030f2ee service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Received unexpected event network-vif-plugged-f2f3c304-8432-48e3-a3ce-70547c912546 for instance with vm_state building and task_state spawning. [ 950.463053] env[63355]: DEBUG nova.network.neutron [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Successfully updated port: f2f3c304-8432-48e3-a3ce-70547c912546 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.496163] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.496677] env[63355]: DEBUG nova.compute.manager [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 950.500063] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.541s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.501642] env[63355]: INFO nova.compute.claims [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.633431] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.730739] env[63355]: DEBUG oslo_vmware.api [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349833, 'name': PowerOnVM_Task, 'duration_secs': 0.694265} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.730888] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 950.731192] env[63355]: INFO nova.compute.manager [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Took 9.55 seconds to spawn the instance on the hypervisor. [ 950.731383] env[63355]: DEBUG nova.compute.manager [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.732231] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be8ca5d-fa87-4701-bea7-71bfceff0321 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.765627] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.846457] env[63355]: DEBUG oslo_vmware.api [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349836, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.966632] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "refresh_cache-fc6a99e6-5319-47d3-8175-770c06d3e325" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.966911] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "refresh_cache-fc6a99e6-5319-47d3-8175-770c06d3e325" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.967130] env[63355]: DEBUG nova.network.neutron [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.006496] env[63355]: DEBUG nova.compute.utils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 951.010287] env[63355]: DEBUG nova.compute.manager [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 951.010445] env[63355]: DEBUG nova.network.neutron [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 951.052483] env[63355]: DEBUG nova.policy [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7bd1307924224936a7df31d9a3538d4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1efb1730acbf4ba5a19bd150bf99d48f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.132896] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.251032] env[63355]: INFO nova.compute.manager [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Took 30.60 seconds to build instance. [ 951.268531] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349835, 'name': ReconfigVM_Task, 'duration_secs': 0.824167} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.268939] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 1570cb36-76e0-4d06-8080-735b5246e92e/1570cb36-76e0-4d06-8080-735b5246e92e.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.269681] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3af534a3-378d-4e36-acce-19df0835da75 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.279213] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 951.279213] env[63355]: value = "task-1349837" [ 951.279213] env[63355]: _type = "Task" [ 951.279213] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.292172] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349837, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.311310] env[63355]: DEBUG nova.network.neutron [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Successfully created port: a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 951.349119] env[63355]: DEBUG oslo_vmware.api [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Task: {'id': task-1349836, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.636774} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.349390] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.349645] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 951.349863] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 951.350104] env[63355]: INFO nova.compute.manager [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Took 1.69 seconds to destroy the instance on the hypervisor. [ 951.350424] env[63355]: DEBUG oslo.service.loopingcall [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.350797] env[63355]: DEBUG nova.compute.manager [-] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 951.350907] env[63355]: DEBUG nova.network.neutron [-] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 951.510805] env[63355]: DEBUG nova.compute.manager [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 951.514913] env[63355]: DEBUG nova.network.neutron [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 951.633430] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.759295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-305fb9e2-e249-401d-8651-5883c26418ed tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.144s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.791552] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349837, 'name': Rename_Task, 'duration_secs': 0.177097} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.795077] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 951.795544] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf4af88f-037d-4d5c-8be0-6102dbf81763 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.803288] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 951.803288] env[63355]: value = "task-1349838" [ 951.803288] env[63355]: _type = "Task" [ 951.803288] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.814543] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349838, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.820211] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe8e61a-526f-4ce3-9ff6-986dc0761476 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.827841] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84fc9aa-d86c-4277-9797-3f99dab7f171 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.860914] env[63355]: DEBUG nova.network.neutron [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Updating instance_info_cache with network_info: [{"id": "f2f3c304-8432-48e3-a3ce-70547c912546", "address": "fa:16:3e:08:7c:be", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2f3c304-84", "ovs_interfaceid": "f2f3c304-8432-48e3-a3ce-70547c912546", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.863443] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb7393b4-3f06-4b74-9422-2310bdb6f5f9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.872562] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7761fa08-eb3e-42bb-8bca-00c782751e7e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.891295] env[63355]: DEBUG nova.compute.provider_tree [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.135341] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.314600] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349838, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.368566] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "refresh_cache-fc6a99e6-5319-47d3-8175-770c06d3e325" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.368566] env[63355]: DEBUG nova.compute.manager [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Instance network_info: |[{"id": "f2f3c304-8432-48e3-a3ce-70547c912546", "address": "fa:16:3e:08:7c:be", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2f3c304-84", "ovs_interfaceid": "f2f3c304-8432-48e3-a3ce-70547c912546", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 952.368566] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:7c:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2f3c304-8432-48e3-a3ce-70547c912546', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.379329] env[63355]: DEBUG oslo.service.loopingcall [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.379580] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.379861] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc402008-f90c-4bde-8eb0-3a999ec665b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.398486] env[63355]: DEBUG nova.scheduler.client.report [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.406276] env[63355]: DEBUG nova.compute.manager [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Received event network-changed-f2f3c304-8432-48e3-a3ce-70547c912546 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.406276] env[63355]: DEBUG nova.compute.manager [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Refreshing instance network info cache due to event network-changed-f2f3c304-8432-48e3-a3ce-70547c912546. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 952.406276] env[63355]: DEBUG oslo_concurrency.lockutils [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] Acquiring lock "refresh_cache-fc6a99e6-5319-47d3-8175-770c06d3e325" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.406276] env[63355]: DEBUG oslo_concurrency.lockutils [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] Acquired lock "refresh_cache-fc6a99e6-5319-47d3-8175-770c06d3e325" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.406276] env[63355]: DEBUG nova.network.neutron [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Refreshing network info cache for port f2f3c304-8432-48e3-a3ce-70547c912546 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.415146] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.415146] env[63355]: value = "task-1349839" [ 952.415146] env[63355]: _type = "Task" [ 952.415146] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.426039] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349839, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.446610] env[63355]: DEBUG nova.network.neutron [-] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.521982] env[63355]: DEBUG nova.compute.manager [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 952.542223] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 952.542521] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 952.542683] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.542868] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 952.543035] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.543195] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 952.543491] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 952.543713] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 952.543923] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 952.544113] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 952.544296] env[63355]: DEBUG nova.virt.hardware [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.545569] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c340a57-d978-4ef8-962a-0c97414c6485 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.553952] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3ff720-b778-4406-b4cc-b3936c985b75 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.635627] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.814252] env[63355]: DEBUG oslo_vmware.api [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349838, 'name': PowerOnVM_Task, 'duration_secs': 0.532404} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.814537] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.814752] env[63355]: INFO nova.compute.manager [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Took 9.19 seconds to spawn the instance on the hypervisor. [ 952.814935] env[63355]: DEBUG nova.compute.manager [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.815759] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f9be99-3468-423b-8b1f-2382b876a862 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.909263] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.909824] env[63355]: DEBUG nova.compute.manager [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 952.914157] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.914157] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.914157] env[63355]: INFO nova.compute.manager [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Shelving [ 952.916032] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.189s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.916440] env[63355]: DEBUG nova.objects.instance [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lazy-loading 'resources' on Instance uuid 74071e73-10fa-4dcb-aa15-91303b2278fb {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.931208] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349839, 'name': CreateVM_Task, 'duration_secs': 0.449338} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.931557] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 952.932696] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.932872] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.933301] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.933892] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0035c0b-a756-465a-bafb-283d85e1ead4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.940889] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 952.940889] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526c50ce-0635-b66a-8fce-7fd11876b220" [ 952.940889] env[63355]: _type = "Task" [ 952.940889] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.950359] env[63355]: INFO nova.compute.manager [-] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Took 1.60 seconds to deallocate network for instance. [ 952.951041] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526c50ce-0635-b66a-8fce-7fd11876b220, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.130744] env[63355]: DEBUG nova.network.neutron [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Successfully updated port: a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.138944] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.216669] env[63355]: DEBUG nova.network.neutron [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Updated VIF entry in instance network info cache for port f2f3c304-8432-48e3-a3ce-70547c912546. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.217098] env[63355]: DEBUG nova.network.neutron [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Updating instance_info_cache with network_info: [{"id": "f2f3c304-8432-48e3-a3ce-70547c912546", "address": "fa:16:3e:08:7c:be", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2f3c304-84", "ovs_interfaceid": "f2f3c304-8432-48e3-a3ce-70547c912546", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.338659] env[63355]: INFO nova.compute.manager [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Took 29.94 seconds to build instance. [ 953.422262] env[63355]: DEBUG nova.compute.utils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.426036] env[63355]: DEBUG nova.compute.manager [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.426036] env[63355]: DEBUG nova.network.neutron [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 953.430244] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.430546] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd4a9bfe-40a3-4438-9a03-cfc343c66eb9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.440377] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 953.440377] env[63355]: value = "task-1349840" [ 953.440377] env[63355]: _type = "Task" [ 953.440377] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.454173] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349840, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.459035] env[63355]: DEBUG oslo_concurrency.lockutils [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.459264] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526c50ce-0635-b66a-8fce-7fd11876b220, 'name': SearchDatastore_Task, 'duration_secs': 0.023495} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.459545] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.459782] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.460028] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.460188] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.460373] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.460656] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c54a9ab-b6c0-449b-a9f8-52e18fb561fa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.469023] env[63355]: DEBUG nova.policy [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7123da3ee74b497395f1681e460aaaca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b6963b8e16b4986a4545914b75a38ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 953.471909] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.472110] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.472847] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f3a9a0d-39aa-4baa-927a-c7a28ad50fd0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.481224] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 953.481224] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52270cbf-2e4d-5a5e-ec0d-722774dcb484" [ 953.481224] env[63355]: _type = "Task" [ 953.481224] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.491037] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52270cbf-2e4d-5a5e-ec0d-722774dcb484, 'name': SearchDatastore_Task, 'duration_secs': 0.009945} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.491640] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc26d558-f915-4544-a9ce-d97b8875ed04 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.497921] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 953.497921] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e5650e-6ef5-0fab-80f1-3c3073b1b957" [ 953.497921] env[63355]: _type = "Task" [ 953.497921] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.505971] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e5650e-6ef5-0fab-80f1-3c3073b1b957, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.638663] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.638830] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.639000] env[63355]: DEBUG nova.network.neutron [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 953.640200] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.720638] env[63355]: DEBUG oslo_concurrency.lockutils [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] Releasing lock "refresh_cache-fc6a99e6-5319-47d3-8175-770c06d3e325" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.720914] env[63355]: DEBUG nova.compute.manager [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Received event network-vif-deleted-2f380960-ecf1-407f-aaf0-88703f8132c3 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.721150] env[63355]: INFO nova.compute.manager [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Neutron deleted interface 2f380960-ecf1-407f-aaf0-88703f8132c3; detaching it from the instance and deleting it from the info cache [ 953.721675] env[63355]: DEBUG nova.network.neutron [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Updating instance_info_cache with network_info: [{"id": "cf363065-5623-4cf2-9d11-fcc3a234909a", "address": "fa:16:3e:fa:92:45", "network": {"id": "1032ad7b-9ec7-4d0e-bdd1-2f0f2e5d82f8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-908697110", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "ad0fba7800a641da9114e93e18d7977d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf363065-56", "ovs_interfaceid": "cf363065-5623-4cf2-9d11-fcc3a234909a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.738240] env[63355]: DEBUG nova.network.neutron [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Successfully created port: 72ce124a-30e1-4794-8c3f-3156b6ce619c {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 953.746018] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-096cc913-b352-4783-b6e2-6946c44d6e6b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.753770] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec570cc-7867-43a5-8699-8f9e63b9115e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.785562] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190f6631-2424-40db-bd3f-69659bae751a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.793702] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014067a0-183c-4314-86db-a07c31fedb33 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.807316] env[63355]: DEBUG nova.compute.provider_tree [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.809284] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.809528] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.809729] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.809908] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.810079] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.811929] env[63355]: INFO nova.compute.manager [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Terminating instance [ 953.813538] env[63355]: DEBUG nova.compute.manager [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 953.813722] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.815024] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1058547f-ecc1-4275-bf05-fa93fcc29b01 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.823213] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.823509] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce6f1420-b34e-432b-ab6d-93ef79098b48 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.831045] env[63355]: DEBUG oslo_vmware.api [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 953.831045] env[63355]: value = "task-1349841" [ 953.831045] env[63355]: _type = "Task" [ 953.831045] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.839367] env[63355]: DEBUG oslo_vmware.api [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349841, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.840850] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ab0e979e-4fe5-4375-8649-49cfa90f5708 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "1570cb36-76e0-4d06-8080-735b5246e92e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.921s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.928938] env[63355]: DEBUG nova.compute.manager [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 953.955023] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349840, 'name': PowerOffVM_Task, 'duration_secs': 0.27543} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.955023] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.955023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535c9ba1-2ae6-43d7-a81a-94af8e4382c4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.979982] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed3e9f3-1540-44f2-a034-2253df27c939 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.011611] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e5650e-6ef5-0fab-80f1-3c3073b1b957, 'name': SearchDatastore_Task, 'duration_secs': 0.013137} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.012160] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.012647] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] fc6a99e6-5319-47d3-8175-770c06d3e325/fc6a99e6-5319-47d3-8175-770c06d3e325.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.013182] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c7fa860-e0a4-49e2-a00a-d64d2363b34f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.026524] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 954.026524] env[63355]: value = "task-1349842" [ 954.026524] env[63355]: _type = "Task" [ 954.026524] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.037744] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349842, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.137136] env[63355]: DEBUG oslo_vmware.api [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349827, 'name': ReconfigVM_Task, 'duration_secs': 5.89803} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.137458] env[63355]: DEBUG oslo_concurrency.lockutils [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.137706] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Reconfigured VM to detach interface {{(pid=63355) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 954.190499] env[63355]: DEBUG nova.network.neutron [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 954.224048] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-142bb614-e728-4c1e-8d8e-8b2ab30e06df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.235232] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f421fda2-619a-4b20-82d0-0fa721d83a1c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.281060] env[63355]: DEBUG nova.compute.manager [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Detach interface failed, port_id=2f380960-ecf1-407f-aaf0-88703f8132c3, reason: Instance 6706c65c-3959-471b-82c0-2196f4aa1ab0 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 954.281549] env[63355]: DEBUG nova.compute.manager [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Received event network-vif-deleted-cf363065-5623-4cf2-9d11-fcc3a234909a {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.281939] env[63355]: INFO nova.compute.manager [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Neutron deleted interface cf363065-5623-4cf2-9d11-fcc3a234909a; detaching it from the instance and deleting it from the info cache [ 954.282279] env[63355]: DEBUG nova.network.neutron [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.310941] env[63355]: DEBUG nova.scheduler.client.report [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.343851] env[63355]: DEBUG oslo_vmware.api [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349841, 'name': PowerOffVM_Task, 'duration_secs': 0.350379} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.346513] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.346630] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.347220] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd68f4e7-76b6-4228-9155-bc37080b6798 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.387913] env[63355]: DEBUG nova.network.neutron [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Updating instance_info_cache with network_info: [{"id": "a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a", "address": "fa:16:3e:f8:de:e7", "network": {"id": "26b0dabd-e662-4b47-bcd4-a50e05f0a52d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-825312565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1efb1730acbf4ba5a19bd150bf99d48f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0ceb2ca-35", "ovs_interfaceid": "a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.442349] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.442577] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.442762] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Deleting the datastore file [datastore2] b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.443397] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7cd898b-a276-41e6-a4a0-5146ecde409c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.453025] env[63355]: DEBUG oslo_vmware.api [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 954.453025] env[63355]: value = "task-1349844" [ 954.453025] env[63355]: _type = "Task" [ 954.453025] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.464870] env[63355]: DEBUG oslo_vmware.api [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.493248] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 954.494134] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-917efa28-fddf-45dc-b499-b4f2f05614f2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.504749] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 954.504749] env[63355]: value = "task-1349845" [ 954.504749] env[63355]: _type = "Task" [ 954.504749] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.516819] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349845, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.542991] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349842, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.607972] env[63355]: DEBUG nova.compute.manager [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Received event network-vif-plugged-a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.608268] env[63355]: DEBUG oslo_concurrency.lockutils [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] Acquiring lock "8e9ff9b7-636f-48df-9168-509d733278f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.608598] env[63355]: DEBUG oslo_concurrency.lockutils [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] Lock "8e9ff9b7-636f-48df-9168-509d733278f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.608847] env[63355]: DEBUG oslo_concurrency.lockutils [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] Lock "8e9ff9b7-636f-48df-9168-509d733278f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.609107] env[63355]: DEBUG nova.compute.manager [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] No waiting events found dispatching network-vif-plugged-a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 954.609373] env[63355]: WARNING nova.compute.manager [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Received unexpected event network-vif-plugged-a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a for instance with vm_state building and task_state spawning. [ 954.609668] env[63355]: DEBUG nova.compute.manager [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Received event network-changed-a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.609898] env[63355]: DEBUG nova.compute.manager [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Refreshing instance network info cache due to event network-changed-a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 954.610142] env[63355]: DEBUG oslo_concurrency.lockutils [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] Acquiring lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.785609] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8adac9ba-0e05-440c-997f-a4f887a8a544 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.795800] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0376f1ae-4820-4c20-82de-56d7626b58fd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.816135] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.819148] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.024s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.819148] env[63355]: DEBUG nova.objects.instance [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lazy-loading 'resources' on Instance uuid dc56b350-cee7-49c4-9712-8c8c29146ff2 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.828566] env[63355]: DEBUG nova.compute.manager [req-232f2261-ac3f-4bea-85e4-73a64cd35340 req-93082b9c-fffc-4eb6-8e75-0fda69d37f6e service nova] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Detach interface failed, port_id=cf363065-5623-4cf2-9d11-fcc3a234909a, reason: Instance 6706c65c-3959-471b-82c0-2196f4aa1ab0 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 954.840427] env[63355]: INFO nova.scheduler.client.report [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Deleted allocations for instance 74071e73-10fa-4dcb-aa15-91303b2278fb [ 954.894561] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.894561] env[63355]: DEBUG nova.compute.manager [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Instance network_info: |[{"id": "a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a", "address": "fa:16:3e:f8:de:e7", "network": {"id": "26b0dabd-e662-4b47-bcd4-a50e05f0a52d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-825312565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1efb1730acbf4ba5a19bd150bf99d48f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0ceb2ca-35", "ovs_interfaceid": "a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 954.894955] env[63355]: DEBUG oslo_concurrency.lockutils [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] Acquired lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.895223] env[63355]: DEBUG nova.network.neutron [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Refreshing network info cache for port a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 954.897442] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:de:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd5970ab5-34b8-4065-bfa6-f568b8f103b7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.905416] env[63355]: DEBUG oslo.service.loopingcall [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.908984] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 954.909617] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-662f78f1-38b5-4107-bad6-a3b5fd66a206 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.931869] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.931869] env[63355]: value = "task-1349846" [ 954.931869] env[63355]: _type = "Task" [ 954.931869] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.946414] env[63355]: DEBUG nova.compute.manager [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 954.950573] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349846, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.966577] env[63355]: DEBUG oslo_vmware.api [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349844, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.31934} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.968919] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.969209] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.969418] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.969646] env[63355]: INFO nova.compute.manager [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Took 1.16 seconds to destroy the instance on the hypervisor. [ 954.969950] env[63355]: DEBUG oslo.service.loopingcall [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.970645] env[63355]: DEBUG nova.compute.manager [-] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 954.970767] env[63355]: DEBUG nova.network.neutron [-] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 954.977152] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='1d5d1c2e864b8b30b6043362da16a332',container_format='bare',created_at=2024-10-24T02:30:07Z,direct_url=,disk_format='vmdk',id=f755e516-c1d3-4618-aedf-9caa138936ff,min_disk=1,min_ram=0,name='tempest-test-snap-2053765567',owner='3b6963b8e16b4986a4545914b75a38ae',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-24T02:30:24Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 954.978058] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 954.978058] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.978058] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 954.978058] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.979178] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 954.979178] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 954.979178] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 954.979178] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 954.979178] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 954.979460] env[63355]: DEBUG nova.virt.hardware [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.980532] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493423b2-5057-4447-a015-256e748a908f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.992130] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfb4d6f-a233-4735-afb7-a286aa975ea2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.026725] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349845, 'name': CreateSnapshot_Task, 'duration_secs': 0.483614} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.026725] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 955.030714] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f819da-912e-41d8-965c-432a7e5af698 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.043068] env[63355]: INFO nova.compute.manager [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Rebuilding instance [ 955.047474] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349842, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.689352} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.048483] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] fc6a99e6-5319-47d3-8175-770c06d3e325/fc6a99e6-5319-47d3-8175-770c06d3e325.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 955.048729] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.048953] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87b9a936-781e-4577-9a17-c8535bac87b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.059634] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 955.059634] env[63355]: value = "task-1349847" [ 955.059634] env[63355]: _type = "Task" [ 955.059634] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.069701] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349847, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.103201] env[63355]: DEBUG nova.compute.manager [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.104127] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872e1303-0e59-4ca0-b520-089d211ead2d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.318689] env[63355]: DEBUG nova.compute.manager [req-b3d6dfc7-41b1-453c-a153-0f1729963a9f req-20b7f505-ac39-4dba-89d5-7951798f1b79 service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Received event network-vif-deleted-ea0634b4-a639-4a82-9851-650958e3eff8 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.318689] env[63355]: INFO nova.compute.manager [req-b3d6dfc7-41b1-453c-a153-0f1729963a9f req-20b7f505-ac39-4dba-89d5-7951798f1b79 service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Neutron deleted interface ea0634b4-a639-4a82-9851-650958e3eff8; detaching it from the instance and deleting it from the info cache [ 955.319127] env[63355]: DEBUG nova.network.neutron [req-b3d6dfc7-41b1-453c-a153-0f1729963a9f req-20b7f505-ac39-4dba-89d5-7951798f1b79 service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.356161] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7066946e-e119-4ab6-b1e4-e35e94c7286d tempest-SecurityGroupsTestJSON-464317749 tempest-SecurityGroupsTestJSON-464317749-project-member] Lock "74071e73-10fa-4dcb-aa15-91303b2278fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.579s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.359498] env[63355]: DEBUG nova.network.neutron [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Updated VIF entry in instance network info cache for port a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 955.360231] env[63355]: DEBUG nova.network.neutron [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Updating instance_info_cache with network_info: [{"id": "a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a", "address": "fa:16:3e:f8:de:e7", "network": {"id": "26b0dabd-e662-4b47-bcd4-a50e05f0a52d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-825312565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1efb1730acbf4ba5a19bd150bf99d48f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0ceb2ca-35", "ovs_interfaceid": "a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.442628] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349846, 'name': CreateVM_Task, 'duration_secs': 0.398605} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.445304] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 955.446224] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.446428] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.446702] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 955.446966] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a13c61fa-6fd0-4986-9b36-b5ceb5c99c73 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.452938] env[63355]: DEBUG oslo_concurrency.lockutils [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.453123] env[63355]: DEBUG oslo_concurrency.lockutils [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.453303] env[63355]: DEBUG nova.network.neutron [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 955.455505] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 955.455505] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5214014d-53cb-62e5-4d69-2c98fe10359a" [ 955.455505] env[63355]: _type = "Task" [ 955.455505] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.467539] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5214014d-53cb-62e5-4d69-2c98fe10359a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.555420] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 955.558490] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-027ba0b4-b5f7-4573-94c0-78eb4cca3e43 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.576207] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349847, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072262} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.576531] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 955.576531] env[63355]: value = "task-1349848" [ 955.576531] env[63355]: _type = "Task" [ 955.576531] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.579679] env[63355]: DEBUG nova.network.neutron [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Successfully updated port: 72ce124a-30e1-4794-8c3f-3156b6ce619c {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.580873] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.582477] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b11c45-ccf7-4f47-a71d-bc30a9ddfdef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.605032] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349848, 'name': CloneVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.613638] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] fc6a99e6-5319-47d3-8175-770c06d3e325/fc6a99e6-5319-47d3-8175-770c06d3e325.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.617011] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0fed0a71-767e-4f8e-b11a-a944d94c4135 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.633390] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.634693] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a69ae22-5fc0-4b49-b257-3449361e3f3e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.642852] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 955.642852] env[63355]: value = "task-1349849" [ 955.642852] env[63355]: _type = "Task" [ 955.642852] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.644304] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 955.644304] env[63355]: value = "task-1349850" [ 955.644304] env[63355]: _type = "Task" [ 955.644304] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.659693] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349849, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.662882] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349850, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.700073] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84626de9-02d3-4c8e-90de-4a9d488a071c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.712236] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b84b9b2-2a6d-4d0f-8929-271c66ae8874 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.750686] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b011dee8-b2bb-4c35-a89a-b72e48a87815 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.759106] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ffd3463-b39b-4e79-95fd-af8d7e1081c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.774528] env[63355]: DEBUG nova.network.neutron [-] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.776295] env[63355]: DEBUG nova.compute.provider_tree [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.822810] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e3cfe1c-fed9-4a57-8c7a-d286700d7ece {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.835472] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0121eb2-d881-40c7-8173-1ffe84e90d9f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.869885] env[63355]: DEBUG oslo_concurrency.lockutils [req-19e2a382-2e89-4a9a-b6ad-63f60f3b6cc9 req-c4a96c7f-94dc-43df-8b9c-1cc22c3b632c service nova] Releasing lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.870463] env[63355]: DEBUG nova.compute.manager [req-b3d6dfc7-41b1-453c-a153-0f1729963a9f req-20b7f505-ac39-4dba-89d5-7951798f1b79 service nova] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Detach interface failed, port_id=ea0634b4-a639-4a82-9851-650958e3eff8, reason: Instance b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 955.971750] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5214014d-53cb-62e5-4d69-2c98fe10359a, 'name': SearchDatastore_Task, 'duration_secs': 0.038512} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.972077] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.972313] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.973137] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.973137] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.973137] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.973266] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9436f74-e8e2-46ac-86a5-e99a7bf6d364 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.982985] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.982985] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 955.983406] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cd177bd-a3b1-4a5b-9007-af0efb6f46de {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.989746] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 955.989746] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ca5294-efe6-13f2-79c3-8180f960e601" [ 955.989746] env[63355]: _type = "Task" [ 955.989746] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.996997] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "386e847e-967b-4247-9730-cdc5ac251474" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.997261] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "386e847e-967b-4247-9730-cdc5ac251474" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.997470] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "386e847e-967b-4247-9730-cdc5ac251474-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.997654] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "386e847e-967b-4247-9730-cdc5ac251474-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.997825] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "386e847e-967b-4247-9730-cdc5ac251474-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.000184] env[63355]: INFO nova.compute.manager [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Terminating instance [ 956.005496] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ca5294-efe6-13f2-79c3-8180f960e601, 'name': SearchDatastore_Task, 'duration_secs': 0.009739} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.006166] env[63355]: DEBUG nova.compute.manager [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 956.006365] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.007638] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5cc5c4-b0c4-4ce4-8e80-c59b293c8369 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.010258] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e74c1de-2886-47e5-80d1-403248a4a097 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.016771] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 956.016771] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523cc7a3-b45d-cab1-8f71-660c58e51b15" [ 956.016771] env[63355]: _type = "Task" [ 956.016771] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.019150] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.022084] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d86cb643-e453-4018-a3f4-c67d64f28ab1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.031825] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523cc7a3-b45d-cab1-8f71-660c58e51b15, 'name': SearchDatastore_Task, 'duration_secs': 0.01065} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.033084] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.033381] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 8e9ff9b7-636f-48df-9168-509d733278f8/8e9ff9b7-636f-48df-9168-509d733278f8.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 956.033724] env[63355]: DEBUG oslo_vmware.api [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 956.033724] env[63355]: value = "task-1349851" [ 956.033724] env[63355]: _type = "Task" [ 956.033724] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.033958] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-752feb6c-25af-4824-a47e-c1bcd6688c51 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.045463] env[63355]: DEBUG oslo_vmware.api [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349851, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.049857] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 956.049857] env[63355]: value = "task-1349852" [ 956.049857] env[63355]: _type = "Task" [ 956.049857] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.059163] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349852, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.082216] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "refresh_cache-248ec2f2-3a41-444a-b550-71c5438a45f7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.082610] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "refresh_cache-248ec2f2-3a41-444a-b550-71c5438a45f7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.082684] env[63355]: DEBUG nova.network.neutron [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 956.096590] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349848, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.159822] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349849, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.166170] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349850, 'name': PowerOffVM_Task, 'duration_secs': 0.254293} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.166554] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.166840] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.168248] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145ac43d-c4f1-4942-b18a-2f52b0b126a6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.177801] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.178274] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8108326f-9f6c-45e3-b3eb-fc4e6a92279e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.247828] env[63355]: INFO nova.network.neutron [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Port 678307c8-eb2a-409d-a339-e40d5fd2ee50 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 956.248053] env[63355]: INFO nova.network.neutron [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Port bd4c85c2-7841-4da0-ba95-43109bfd107e from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 956.248408] env[63355]: DEBUG nova.network.neutron [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [{"id": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "address": "fa:16:3e:a9:13:18", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348dcf0-b9", "ovs_interfaceid": "2348dcf0-b9a5-491f-87cb-a14e9e002368", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.279042] env[63355]: INFO nova.compute.manager [-] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Took 1.31 seconds to deallocate network for instance. [ 956.279830] env[63355]: DEBUG nova.scheduler.client.report [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.400782] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.400782] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.400782] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleting the datastore file [datastore1] 1570cb36-76e0-4d06-8080-735b5246e92e {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.401698] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46b3064b-944c-4505-a3d8-f1821dd6c16f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.414040] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 956.414040] env[63355]: value = "task-1349854" [ 956.414040] env[63355]: _type = "Task" [ 956.414040] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.425326] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349854, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.548443] env[63355]: DEBUG oslo_vmware.api [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349851, 'name': PowerOffVM_Task, 'duration_secs': 0.240747} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.548756] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.548954] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.549294] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a1a7f22-3b74-4b36-9846-e8d0522b389e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.561247] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349852, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466334} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.561679] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 8e9ff9b7-636f-48df-9168-509d733278f8/8e9ff9b7-636f-48df-9168-509d733278f8.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.561801] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.562295] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-42bd8357-d7c9-4d80-88a3-95a48b5cbe9a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.570238] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 956.570238] env[63355]: value = "task-1349856" [ 956.570238] env[63355]: _type = "Task" [ 956.570238] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.581066] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349856, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.593474] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349848, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.624223] env[63355]: DEBUG nova.network.neutron [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.647038] env[63355]: DEBUG nova.compute.manager [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received event network-vif-deleted-678307c8-eb2a-409d-a339-e40d5fd2ee50 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.647336] env[63355]: DEBUG nova.compute.manager [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Received event network-vif-plugged-72ce124a-30e1-4794-8c3f-3156b6ce619c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.647603] env[63355]: DEBUG oslo_concurrency.lockutils [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] Acquiring lock "248ec2f2-3a41-444a-b550-71c5438a45f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.647734] env[63355]: DEBUG oslo_concurrency.lockutils [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] Lock "248ec2f2-3a41-444a-b550-71c5438a45f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.647907] env[63355]: DEBUG oslo_concurrency.lockutils [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] Lock "248ec2f2-3a41-444a-b550-71c5438a45f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.648086] env[63355]: DEBUG nova.compute.manager [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] No waiting events found dispatching network-vif-plugged-72ce124a-30e1-4794-8c3f-3156b6ce619c {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 956.648265] env[63355]: WARNING nova.compute.manager [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Received unexpected event network-vif-plugged-72ce124a-30e1-4794-8c3f-3156b6ce619c for instance with vm_state building and task_state spawning. [ 956.648420] env[63355]: DEBUG nova.compute.manager [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received event network-vif-deleted-bd4c85c2-7841-4da0-ba95-43109bfd107e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.648588] env[63355]: DEBUG nova.compute.manager [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Received event network-changed-72ce124a-30e1-4794-8c3f-3156b6ce619c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.648742] env[63355]: DEBUG nova.compute.manager [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Refreshing instance network info cache due to event network-changed-72ce124a-30e1-4794-8c3f-3156b6ce619c. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 956.648913] env[63355]: DEBUG oslo_concurrency.lockutils [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] Acquiring lock "refresh_cache-248ec2f2-3a41-444a-b550-71c5438a45f7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.659615] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349849, 'name': ReconfigVM_Task, 'duration_secs': 0.839069} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.659891] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Reconfigured VM instance instance-00000050 to attach disk [datastore1] fc6a99e6-5319-47d3-8175-770c06d3e325/fc6a99e6-5319-47d3-8175-770c06d3e325.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.660523] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2716c618-666a-4c3a-a111-6789515c5598 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.672054] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 956.672054] env[63355]: value = "task-1349857" [ 956.672054] env[63355]: _type = "Task" [ 956.672054] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.681717] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349857, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.695534] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.695778] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.696182] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Deleting the datastore file [datastore2] 386e847e-967b-4247-9730-cdc5ac251474 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.696413] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3869b61-3bd0-458d-b529-4044e4db9b72 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.703992] env[63355]: DEBUG oslo_vmware.api [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 956.703992] env[63355]: value = "task-1349858" [ 956.703992] env[63355]: _type = "Task" [ 956.703992] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.712694] env[63355]: DEBUG oslo_vmware.api [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349858, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.751848] env[63355]: DEBUG oslo_concurrency.lockutils [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-386e847e-967b-4247-9730-cdc5ac251474" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.768990] env[63355]: DEBUG nova.network.neutron [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Updating instance_info_cache with network_info: [{"id": "72ce124a-30e1-4794-8c3f-3156b6ce619c", "address": "fa:16:3e:7e:7b:60", "network": {"id": "60046874-e6ff-46c0-9843-752048f98e4a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1700834300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b6963b8e16b4986a4545914b75a38ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72ce124a-30", "ovs_interfaceid": "72ce124a-30e1-4794-8c3f-3156b6ce619c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.792661] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.974s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.795021] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.796032] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 20.053s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.822495] env[63355]: INFO nova.scheduler.client.report [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Deleted allocations for instance dc56b350-cee7-49c4-9712-8c8c29146ff2 [ 956.922731] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349854, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.080751] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349856, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.317356} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.081062] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.081860] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb96ddbb-4e58-4a3c-858a-af66de07d25e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.093427] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349848, 'name': CloneVM_Task, 'duration_secs': 1.480458} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.102431] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Created linked-clone VM from snapshot [ 957.111282] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 8e9ff9b7-636f-48df-9168-509d733278f8/8e9ff9b7-636f-48df-9168-509d733278f8.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.112098] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4a09d8-3e89-4658-8fd5-beffd306596c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.114814] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab1d60e6-e09c-463d-9ed8-31479bcb36f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.135340] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Uploading image 95245748-2b88-4a4c-8d0b-6a828a676666 {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 957.139027] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 957.139027] env[63355]: value = "task-1349859" [ 957.139027] env[63355]: _type = "Task" [ 957.139027] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.148033] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349859, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.163662] env[63355]: DEBUG oslo_vmware.rw_handles [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 957.163662] env[63355]: value = "vm-287710" [ 957.163662] env[63355]: _type = "VirtualMachine" [ 957.163662] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 957.163843] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7f94bdfc-88fe-417e-a2b0-4a0e06c82f82 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.172697] env[63355]: DEBUG oslo_vmware.rw_handles [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lease: (returnval){ [ 957.172697] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c81d10-a12e-7871-f18a-71630feeab56" [ 957.172697] env[63355]: _type = "HttpNfcLease" [ 957.172697] env[63355]: } obtained for exporting VM: (result){ [ 957.172697] env[63355]: value = "vm-287710" [ 957.172697] env[63355]: _type = "VirtualMachine" [ 957.172697] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 957.173397] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the lease: (returnval){ [ 957.173397] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c81d10-a12e-7871-f18a-71630feeab56" [ 957.173397] env[63355]: _type = "HttpNfcLease" [ 957.173397] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 957.188146] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349857, 'name': Rename_Task, 'duration_secs': 0.249735} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.188418] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 957.188418] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c81d10-a12e-7871-f18a-71630feeab56" [ 957.188418] env[63355]: _type = "HttpNfcLease" [ 957.188418] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 957.188703] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.188953] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52a70375-a69a-4af1-bba6-51f1519fd01c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.196661] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 957.196661] env[63355]: value = "task-1349861" [ 957.196661] env[63355]: _type = "Task" [ 957.196661] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.207131] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349861, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.215333] env[63355]: DEBUG oslo_vmware.api [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349858, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204238} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.215613] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.215846] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.216053] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.216295] env[63355]: INFO nova.compute.manager [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Took 1.21 seconds to destroy the instance on the hypervisor. [ 957.216565] env[63355]: DEBUG oslo.service.loopingcall [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.216737] env[63355]: DEBUG nova.compute.manager [-] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 957.216829] env[63355]: DEBUG nova.network.neutron [-] [instance: 386e847e-967b-4247-9730-cdc5ac251474] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 957.256576] env[63355]: DEBUG oslo_concurrency.lockutils [None req-687ebb5e-c769-4c9b-9b70-d23a772c8fa3 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-386e847e-967b-4247-9730-cdc5ac251474-678307c8-eb2a-409d-a339-e40d5fd2ee50" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.753s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.265208] env[63355]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port bd4c85c2-7841-4da0-ba95-43109bfd107e could not be found.", "detail": ""}} {{(pid=63355) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 957.265208] env[63355]: DEBUG nova.network.neutron [-] Unable to show port bd4c85c2-7841-4da0-ba95-43109bfd107e as it no longer exists. {{(pid=63355) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 957.273879] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "refresh_cache-248ec2f2-3a41-444a-b550-71c5438a45f7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.273879] env[63355]: DEBUG nova.compute.manager [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Instance network_info: |[{"id": "72ce124a-30e1-4794-8c3f-3156b6ce619c", "address": "fa:16:3e:7e:7b:60", "network": {"id": "60046874-e6ff-46c0-9843-752048f98e4a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1700834300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b6963b8e16b4986a4545914b75a38ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72ce124a-30", "ovs_interfaceid": "72ce124a-30e1-4794-8c3f-3156b6ce619c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 957.273879] env[63355]: DEBUG oslo_concurrency.lockutils [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] Acquired lock "refresh_cache-248ec2f2-3a41-444a-b550-71c5438a45f7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.273879] env[63355]: DEBUG nova.network.neutron [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Refreshing network info cache for port 72ce124a-30e1-4794-8c3f-3156b6ce619c {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 957.273879] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:7b:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'df1bf911-aac9-4d2d-ae69-66ace3e6a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72ce124a-30e1-4794-8c3f-3156b6ce619c', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.286044] env[63355]: DEBUG oslo.service.loopingcall [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.286044] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 957.286044] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b75cbd5a-61c3-412b-8693-fb3b2510f2f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.311560] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.311560] env[63355]: value = "task-1349862" [ 957.311560] env[63355]: _type = "Task" [ 957.311560] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.321405] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349862, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.334306] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3ea7f880-ec51-4ef5-b4cb-86ed29cc92b8 tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "dc56b350-cee7-49c4-9712-8c8c29146ff2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.577s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.428262] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349854, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.522745} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.428730] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.429069] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.429393] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.649997] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.685503] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 957.685503] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c81d10-a12e-7871-f18a-71630feeab56" [ 957.685503] env[63355]: _type = "HttpNfcLease" [ 957.685503] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 957.685838] env[63355]: DEBUG oslo_vmware.rw_handles [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 957.685838] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c81d10-a12e-7871-f18a-71630feeab56" [ 957.685838] env[63355]: _type = "HttpNfcLease" [ 957.685838] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 957.686737] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce23784-d49f-4a75-b1ea-bbf98989b1df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.695803] env[63355]: DEBUG oslo_vmware.rw_handles [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52278bd2-2895-8cc5-e798-a1b7f9aad49c/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 957.696014] env[63355]: DEBUG oslo_vmware.rw_handles [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52278bd2-2895-8cc5-e798-a1b7f9aad49c/disk-0.vmdk for reading. {{(pid=63355) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 957.775532] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349861, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.822801] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349862, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.825775] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ddbf5df6-04e6-40f8-9b44-2d1ac91cf8cb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.847141] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 386e847e-967b-4247-9730-cdc5ac251474 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.847284] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 0b660647-697a-445b-bca4-82ef6975ea75 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.847513] env[63355]: WARNING nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 957.847729] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance f5e62ce1-40b7-4648-a4a6-068ff06eaf9b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.847882] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8e3ec9d3-bc22-4e39-ad7c-93268dd59020 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.848101] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.848378] env[63355]: WARNING nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 199bc488-2e5d-4cea-aefb-ddd35ecc7a30 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 957.848378] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance ecbacdb3-516e-43c3-96e3-4961b76565ca actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.848836] env[63355]: WARNING nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance d2480bd6-527b-46b8-8a6a-22ad4eda3f5f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 957.849088] env[63355]: WARNING nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance a408e19a-0960-430d-8550-0a304c63da61 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 957.849290] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7c153109-b814-4e11-b4f9-7b8cebb853d4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.849456] env[63355]: WARNING nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 6706c65c-3959-471b-82c0-2196f4aa1ab0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 957.849725] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance cd9de9b8-ad42-4dbb-b435-927738a55f3f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.849725] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 1570cb36-76e0-4d06-8080-735b5246e92e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.849932] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance fc6a99e6-5319-47d3-8175-770c06d3e325 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.850113] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8e9ff9b7-636f-48df-9168-509d733278f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 957.850305] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 248ec2f2-3a41-444a-b550-71c5438a45f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.121287] env[63355]: DEBUG nova.network.neutron [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Updated VIF entry in instance network info cache for port 72ce124a-30e1-4794-8c3f-3156b6ce619c. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 958.121287] env[63355]: DEBUG nova.network.neutron [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Updating instance_info_cache with network_info: [{"id": "72ce124a-30e1-4794-8c3f-3156b6ce619c", "address": "fa:16:3e:7e:7b:60", "network": {"id": "60046874-e6ff-46c0-9843-752048f98e4a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1700834300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b6963b8e16b4986a4545914b75a38ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72ce124a-30", "ovs_interfaceid": "72ce124a-30e1-4794-8c3f-3156b6ce619c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.156137] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349859, 'name': ReconfigVM_Task, 'duration_secs': 0.635396} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.156705] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 8e9ff9b7-636f-48df-9168-509d733278f8/8e9ff9b7-636f-48df-9168-509d733278f8.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.157712] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-958fc369-54df-46b1-b05f-75b3c19ea503 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.166852] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 958.166852] env[63355]: value = "task-1349863" [ 958.166852] env[63355]: _type = "Task" [ 958.166852] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.181945] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349863, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.208924] env[63355]: DEBUG oslo_vmware.api [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349861, 'name': PowerOnVM_Task, 'duration_secs': 0.744717} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.209584] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 958.210024] env[63355]: INFO nova.compute.manager [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Took 8.12 seconds to spawn the instance on the hypervisor. [ 958.210570] env[63355]: DEBUG nova.compute.manager [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.211749] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b3f33c-ec1a-4f3d-b5dc-d47896a4cdd7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.983529] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance ae3961d2-dc5b-4e49-acca-6fb52291f23b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 958.987047] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 958.988306] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2816MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 958.993076] env[63355]: DEBUG nova.network.neutron [-] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.994640] env[63355]: DEBUG oslo_concurrency.lockutils [req-2c1ec26d-795d-407f-9709-28eecbda8d3e req-81ab2269-b682-43ae-a9c9-b6820fef3e01 service nova] Releasing lock "refresh_cache-248ec2f2-3a41-444a-b550-71c5438a45f7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.996199] env[63355]: DEBUG nova.compute.manager [req-f4677328-bc2d-4626-bc3b-7b2498470d56 req-aae2447e-a836-4cda-ae6a-5631770151f8 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Received event network-vif-deleted-2348dcf0-b9a5-491f-87cb-a14e9e002368 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.996750] env[63355]: INFO nova.compute.manager [req-f4677328-bc2d-4626-bc3b-7b2498470d56 req-aae2447e-a836-4cda-ae6a-5631770151f8 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Neutron deleted interface 2348dcf0-b9a5-491f-87cb-a14e9e002368; detaching it from the instance and deleting it from the info cache [ 958.997421] env[63355]: DEBUG nova.network.neutron [req-f4677328-bc2d-4626-bc3b-7b2498470d56 req-aae2447e-a836-4cda-ae6a-5631770151f8 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.020740] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349863, 'name': Rename_Task, 'duration_secs': 0.355312} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.021575] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349862, 'name': CreateVM_Task, 'duration_secs': 0.634351} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.023018] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 959.023079] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 959.023621] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32eb1bd4-c20d-4ab1-9c0d-4ca6f020111d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.026695] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.026740] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.027362] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.030954] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e308d609-9e2a-40f9-8da4-bb23e9991dca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.040594] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 959.040594] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b4187c-f853-d78e-4839-9c4426fedc18" [ 959.040594] env[63355]: _type = "Task" [ 959.040594] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.041780] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 959.041780] env[63355]: value = "task-1349864" [ 959.041780] env[63355]: _type = "Task" [ 959.041780] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.050829] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.054208] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.054208] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.054208] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.054208] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.054208] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.054208] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.054208] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.054208] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.054208] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.054208] env[63355]: DEBUG nova.virt.hardware [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.058053] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342c1258-3f6a-4f58-8cff-13e2cc7f83c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.075751] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0709a7-edd6-4937-bf3b-7363763e897a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.081784] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349864, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.085132] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.085132] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Processing image f755e516-c1d3-4618-aedf-9caa138936ff {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.085132] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff/f755e516-c1d3-4618-aedf-9caa138936ff.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.085132] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff/f755e516-c1d3-4618-aedf-9caa138936ff.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.085132] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.086316] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5da99786-dd16-42e7-a24b-6401dc815fbb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.100036] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:1e:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '130387c4-e4ec-4d95-8e9d-bb079baabad8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91c7c439-7d4c-4512-90c6-d1698630ae2b', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.108303] env[63355]: DEBUG oslo.service.loopingcall [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.114055] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.114055] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.114055] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.114595] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b90651a5-f8df-4ad6-9b9e-a09d887eae54 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.132190] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-925002c5-bf0a-42d6-af9a-97ce6a8022a7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.143114] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 959.143114] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52840307-f4d8-af79-83cc-fcc8f837cd5a" [ 959.143114] env[63355]: _type = "Task" [ 959.143114] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.148944] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.148944] env[63355]: value = "task-1349865" [ 959.148944] env[63355]: _type = "Task" [ 959.148944] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.164674] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349865, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.165114] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Preparing fetch location {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 959.165425] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Fetch image to [datastore2] OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461/OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461.vmdk {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 959.165769] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Downloading stream optimized image f755e516-c1d3-4618-aedf-9caa138936ff to [datastore2] OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461/OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461.vmdk on the data store datastore2 as vApp {{(pid=63355) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 959.166027] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Downloading image file data f755e516-c1d3-4618-aedf-9caa138936ff to the ESX as VM named 'OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461' {{(pid=63355) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 959.269439] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 959.269439] env[63355]: value = "resgroup-9" [ 959.269439] env[63355]: _type = "ResourcePool" [ 959.269439] env[63355]: }. {{(pid=63355) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 959.270634] env[63355]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-240cd68f-d4a6-4fc6-86b8-676d9f7fba18 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.307900] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lease: (returnval){ [ 959.307900] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e0b317-ca21-85ef-e393-6c22d53c6ee6" [ 959.307900] env[63355]: _type = "HttpNfcLease" [ 959.307900] env[63355]: } obtained for vApp import into resource pool (val){ [ 959.307900] env[63355]: value = "resgroup-9" [ 959.307900] env[63355]: _type = "ResourcePool" [ 959.307900] env[63355]: }. {{(pid=63355) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 959.309698] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the lease: (returnval){ [ 959.309698] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e0b317-ca21-85ef-e393-6c22d53c6ee6" [ 959.309698] env[63355]: _type = "HttpNfcLease" [ 959.309698] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 959.316845] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.316845] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e0b317-ca21-85ef-e393-6c22d53c6ee6" [ 959.316845] env[63355]: _type = "HttpNfcLease" [ 959.316845] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 959.430721] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef08ff5-4c2a-4759-9a54-27888ef9caad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.442205] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca4861f-ed7f-4714-9a37-db692c1ecd4b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.475115] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd95d51-f985-48d8-aeff-c8af8012f7ab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.484569] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada6ac35-a32f-4166-99a5-47046b15e0e9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.499979] env[63355]: INFO nova.compute.manager [-] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Took 2.28 seconds to deallocate network for instance. [ 959.500574] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 959.508359] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43f0b2d9-e894-49d9-9ba5-99756fbfab16 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.531998] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabd7758-b0ce-4c8c-bd03-953ee751e22c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.545893] env[63355]: INFO nova.compute.manager [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Took 34.42 seconds to build instance. [ 959.562892] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349864, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.574807] env[63355]: DEBUG nova.compute.manager [req-f4677328-bc2d-4626-bc3b-7b2498470d56 req-aae2447e-a836-4cda-ae6a-5631770151f8 service nova] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Detach interface failed, port_id=2348dcf0-b9a5-491f-87cb-a14e9e002368, reason: Instance 386e847e-967b-4247-9730-cdc5ac251474 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 959.661064] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349865, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.817486] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.817486] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e0b317-ca21-85ef-e393-6c22d53c6ee6" [ 959.817486] env[63355]: _type = "HttpNfcLease" [ 959.817486] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 960.020656] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.030016] env[63355]: ERROR nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [req-99967e48-95f2-452b-9dd2-7423d935f9f1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 47abb610-db7e-4770-911d-187dd075ef8b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-99967e48-95f2-452b-9dd2-7423d935f9f1"}]} [ 960.049231] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Refreshing inventories for resource provider 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 960.055351] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b8c2189a-4e45-400c-8e47-221595570542 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "fc6a99e6-5319-47d3-8175-770c06d3e325" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.585s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.063197] env[63355]: DEBUG oslo_vmware.api [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349864, 'name': PowerOnVM_Task, 'duration_secs': 0.660893} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.063197] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 960.063197] env[63355]: INFO nova.compute.manager [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Took 7.54 seconds to spawn the instance on the hypervisor. [ 960.063362] env[63355]: DEBUG nova.compute.manager [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.064347] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c118a3-a38e-4fe4-bc14-ad4ce52aaa5e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.070479] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Updating ProviderTree inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 960.070585] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 960.090985] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Refreshing aggregate associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, aggregates: None {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 960.109813] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Refreshing trait associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 960.164356] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349865, 'name': CreateVM_Task, 'duration_secs': 0.55628} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.167563] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 960.169569] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.169569] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.169799] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 960.170038] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-438ad2f6-977b-4ebc-b877-2e98d1ed2151 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.175780] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 960.175780] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f07de8-d1ed-0e00-cdc1-441d286cfba8" [ 960.175780] env[63355]: _type = "Task" [ 960.175780] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.185959] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f07de8-d1ed-0e00-cdc1-441d286cfba8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.318956] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.318956] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e0b317-ca21-85ef-e393-6c22d53c6ee6" [ 960.318956] env[63355]: _type = "HttpNfcLease" [ 960.318956] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 960.319893] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 960.319893] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e0b317-ca21-85ef-e393-6c22d53c6ee6" [ 960.319893] env[63355]: _type = "HttpNfcLease" [ 960.319893] env[63355]: }. {{(pid=63355) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 960.322383] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca07897f-f89d-4dc8-aa67-0e897dee4f39 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.330940] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d5df1-6112-0f72-d99b-da45288f1090/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 960.332307] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d5df1-6112-0f72-d99b-da45288f1090/disk-0.vmdk. {{(pid=63355) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 960.404263] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9ee7e4f8-2a26-4aef-9fda-b363989f34a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.442337] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01975005-9c13-4522-91aa-28bd0eb5148e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.452023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6414795c-5ea8-44fd-bc04-fa6582295dde {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.484546] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b555de7-6491-42ef-856c-e10b9837101d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.493411] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34c023c-232c-464d-8b7e-935b52d19ff2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.508939] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 960.588831] env[63355]: INFO nova.compute.manager [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Took 29.96 seconds to build instance. [ 960.688316] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f07de8-d1ed-0e00-cdc1-441d286cfba8, 'name': SearchDatastore_Task, 'duration_secs': 0.011033} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.688698] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.689054] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.689326] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.689476] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.689667] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.690039] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9b05251-3558-4686-89d8-e0c6f9d3dfc5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.700720] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.700935] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 960.701674] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39670544-89b2-4e0c-bf06-3043240e8371 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.707780] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 960.707780] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ae422f-29ff-2ab5-6a3e-658e3eceab26" [ 960.707780] env[63355]: _type = "Task" [ 960.707780] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.716864] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ae422f-29ff-2ab5-6a3e-658e3eceab26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.055569] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Updated inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b with generation 95 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 961.055569] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Updating resource provider 47abb610-db7e-4770-911d-187dd075ef8b generation from 95 to 96 during operation: update_inventory {{(pid=63355) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 961.055893] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 961.095329] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5ff023fa-2d03-4177-abea-4c75f2bcf624 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e9ff9b7-636f-48df-9168-509d733278f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.621s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.128646] env[63355]: DEBUG oslo_concurrency.lockutils [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "fc6a99e6-5319-47d3-8175-770c06d3e325" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.128940] env[63355]: DEBUG oslo_concurrency.lockutils [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "fc6a99e6-5319-47d3-8175-770c06d3e325" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.129181] env[63355]: DEBUG oslo_concurrency.lockutils [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "fc6a99e6-5319-47d3-8175-770c06d3e325-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.129837] env[63355]: DEBUG oslo_concurrency.lockutils [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "fc6a99e6-5319-47d3-8175-770c06d3e325-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.130129] env[63355]: DEBUG oslo_concurrency.lockutils [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "fc6a99e6-5319-47d3-8175-770c06d3e325-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.133318] env[63355]: INFO nova.compute.manager [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Terminating instance [ 961.138664] env[63355]: DEBUG nova.compute.manager [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 961.140117] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.140634] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2ca7d9-2cca-403b-b7e2-d9fb4839f020 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.157846] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 961.160284] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ccc50fa3-4694-454f-9b7a-61409e7c6d2a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.171147] env[63355]: DEBUG oslo_vmware.api [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 961.171147] env[63355]: value = "task-1349867" [ 961.171147] env[63355]: _type = "Task" [ 961.171147] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.183188] env[63355]: DEBUG oslo_vmware.api [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.223387] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ae422f-29ff-2ab5-6a3e-658e3eceab26, 'name': SearchDatastore_Task, 'duration_secs': 0.012406} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.225992] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fc75acb-160e-416a-81cf-b2fecc695ca0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.235328] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 961.235328] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529a7667-76bb-7eda-f5d1-1a35c112d56f" [ 961.235328] env[63355]: _type = "Task" [ 961.235328] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.246513] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529a7667-76bb-7eda-f5d1-1a35c112d56f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.247239] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Completed reading data from the image iterator. {{(pid=63355) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 961.247631] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d5df1-6112-0f72-d99b-da45288f1090/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 961.248625] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24435d66-84a4-4140-ba8d-4fe29de2267a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.261009] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d5df1-6112-0f72-d99b-da45288f1090/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 961.261628] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d5df1-6112-0f72-d99b-da45288f1090/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 961.264248] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9092ed1e-d451-4d41-bf6d-960bf0ae9302 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.566602] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 961.566602] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.769s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.566602] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.718s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.566602] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.569228] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.989s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.569609] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.573106] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.101s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.573106] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.575364] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.006s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.577181] env[63355]: INFO nova.compute.claims [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.611098] env[63355]: INFO nova.scheduler.client.report [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Deleted allocations for instance 199bc488-2e5d-4cea-aefb-ddd35ecc7a30 [ 961.617042] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/529d5df1-6112-0f72-d99b-da45288f1090/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 961.617042] env[63355]: INFO nova.virt.vmwareapi.images [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Downloaded image file data f755e516-c1d3-4618-aedf-9caa138936ff [ 961.618842] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ed300b-96c1-4f58-8f6e-ce512beae944 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.642417] env[63355]: INFO nova.scheduler.client.report [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Deleted allocations for instance d2480bd6-527b-46b8-8a6a-22ad4eda3f5f [ 961.643256] env[63355]: INFO nova.scheduler.client.report [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Deleted allocations for instance a408e19a-0960-430d-8550-0a304c63da61 [ 961.644989] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6666fe3e-7a54-42de-8baf-4aacbf9f6a29 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.683576] env[63355]: DEBUG oslo_vmware.api [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349867, 'name': PowerOffVM_Task, 'duration_secs': 0.248084} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.684564] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 961.685594] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 961.687966] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a455b835-d990-4cb1-9d5c-3325425dcb16 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.690113] env[63355]: INFO nova.virt.vmwareapi.images [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] The imported VM was unregistered [ 961.692706] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Caching image {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 961.693034] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Creating directory with path [datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.694009] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ebb875d-65ab-4606-ac23-ee17ad5a1983 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.709426] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Created directory with path [datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.709675] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461/OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461.vmdk to [datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff/f755e516-c1d3-4618-aedf-9caa138936ff.vmdk. {{(pid=63355) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 961.709958] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-9e6545bb-8674-42be-8c70-c13a05de8a30 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.720365] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 961.720365] env[63355]: value = "task-1349870" [ 961.720365] env[63355]: _type = "Task" [ 961.720365] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.733590] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349870, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.747587] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529a7667-76bb-7eda-f5d1-1a35c112d56f, 'name': SearchDatastore_Task, 'duration_secs': 0.014779} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.747902] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.748230] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 1570cb36-76e0-4d06-8080-735b5246e92e/1570cb36-76e0-4d06-8080-735b5246e92e.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 961.748520] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13448962-0a93-4e97-a186-a4a1e96d4cad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.756764] env[63355]: INFO nova.compute.manager [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Rescuing [ 961.757075] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.757242] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.757421] env[63355]: DEBUG nova.network.neutron [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 961.758738] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 961.758738] env[63355]: value = "task-1349871" [ 961.758738] env[63355]: _type = "Task" [ 961.758738] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.771347] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349871, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.784867] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 961.785043] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 961.785224] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleting the datastore file [datastore1] fc6a99e6-5319-47d3-8175-770c06d3e325 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.785519] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40347334-3219-431e-9b10-2142de333e78 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.796028] env[63355]: DEBUG oslo_vmware.api [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 961.796028] env[63355]: value = "task-1349872" [ 961.796028] env[63355]: _type = "Task" [ 961.796028] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.804911] env[63355]: DEBUG oslo_vmware.api [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349872, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.122161] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e315b193-60cb-41ad-9cf4-150862e0f1fb tempest-MultipleCreateTestJSON-1948953373 tempest-MultipleCreateTestJSON-1948953373-project-member] Lock "199bc488-2e5d-4cea-aefb-ddd35ecc7a30" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.356s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.159732] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f8a28d69-97ac-485b-8f2b-85d59ed1c772 tempest-InstanceActionsV221TestJSON-875825468 tempest-InstanceActionsV221TestJSON-875825468-project-member] Lock "a408e19a-0960-430d-8550-0a304c63da61" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.899s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.160972] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6626c4de-0469-456d-b7e6-6d0bfb750c12 tempest-VolumesAdminNegativeTest-2121238973 tempest-VolumesAdminNegativeTest-2121238973-project-member] Lock "d2480bd6-527b-46b8-8a6a-22ad4eda3f5f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.059s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.234853] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349870, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.278915] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349871, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.312578] env[63355]: DEBUG oslo_vmware.api [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349872, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165605} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.313243] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.313485] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 962.313840] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 962.314253] env[63355]: INFO nova.compute.manager [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Took 1.18 seconds to destroy the instance on the hypervisor. [ 962.314690] env[63355]: DEBUG oslo.service.loopingcall [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.315282] env[63355]: DEBUG nova.compute.manager [-] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.315408] env[63355]: DEBUG nova.network.neutron [-] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 962.738569] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349870, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.772127] env[63355]: DEBUG nova.network.neutron [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Updating instance_info_cache with network_info: [{"id": "a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a", "address": "fa:16:3e:f8:de:e7", "network": {"id": "26b0dabd-e662-4b47-bcd4-a50e05f0a52d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-825312565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1efb1730acbf4ba5a19bd150bf99d48f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0ceb2ca-35", "ovs_interfaceid": "a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.777903] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349871, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.899008] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1e930c-e2cf-44dd-bc85-e6874adadde5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.915818] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a668cc8-d61d-4d03-a140-22a516981590 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.965238] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eba0952-8db2-48fe-aeaf-f873e22a6227 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.981557] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50908902-1156-4ff6-a6a6-e76511c1ea25 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.002954] env[63355]: DEBUG nova.compute.provider_tree [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.236035] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349870, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.273713] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349871, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.489028} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.274111] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 1570cb36-76e0-4d06-8080-735b5246e92e/1570cb36-76e0-4d06-8080-735b5246e92e.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.274691] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.274999] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-16d19550-8b65-4ece-b760-6e8db76d9e04 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.282878] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.304140] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 963.304140] env[63355]: value = "task-1349873" [ 963.304140] env[63355]: _type = "Task" [ 963.304140] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.322740] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349873, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.341159] env[63355]: DEBUG nova.compute.manager [req-1448406d-c309-428d-ab59-8ada0a74b644 req-7902ac6e-c191-49d4-a29d-ac2e08b36526 service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Received event network-vif-deleted-f2f3c304-8432-48e3-a3ce-70547c912546 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.341414] env[63355]: INFO nova.compute.manager [req-1448406d-c309-428d-ab59-8ada0a74b644 req-7902ac6e-c191-49d4-a29d-ac2e08b36526 service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Neutron deleted interface f2f3c304-8432-48e3-a3ce-70547c912546; detaching it from the instance and deleting it from the info cache [ 963.341587] env[63355]: DEBUG nova.network.neutron [req-1448406d-c309-428d-ab59-8ada0a74b644 req-7902ac6e-c191-49d4-a29d-ac2e08b36526 service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.472050] env[63355]: DEBUG nova.network.neutron [-] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.509679] env[63355]: DEBUG nova.scheduler.client.report [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.559282] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquiring lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.559593] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.743335] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349870, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.819931] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349873, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111643} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.820420] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.821634] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e85a7a5-ee59-4d24-b98c-e1d9b93e919d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.830732] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.831483] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01523965-b713-49c3-96b1-81f2b663ea77 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.863599] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 1570cb36-76e0-4d06-8080-735b5246e92e/1570cb36-76e0-4d06-8080-735b5246e92e.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.865512] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b2742149-3ad4-4ac6-86a2-1b33b6f8df1a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.868346] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10a8a4c8-f181-4779-8de0-1e71de9ce935 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.884428] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 963.884428] env[63355]: value = "task-1349874" [ 963.884428] env[63355]: _type = "Task" [ 963.884428] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.898204] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.899138] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 963.899138] env[63355]: value = "task-1349875" [ 963.899138] env[63355]: _type = "Task" [ 963.899138] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.902362] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3d7f3b-44e3-47d9-b3c3-4fea271312d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.935858] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349875, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.946078] env[63355]: DEBUG nova.compute.manager [req-1448406d-c309-428d-ab59-8ada0a74b644 req-7902ac6e-c191-49d4-a29d-ac2e08b36526 service nova] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Detach interface failed, port_id=f2f3c304-8432-48e3-a3ce-70547c912546, reason: Instance fc6a99e6-5319-47d3-8175-770c06d3e325 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 963.972200] env[63355]: INFO nova.compute.manager [-] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Took 1.66 seconds to deallocate network for instance. [ 964.016388] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.017695] env[63355]: DEBUG nova.compute.manager [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 964.020023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.561s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.020253] env[63355]: DEBUG oslo_concurrency.lockutils [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.022732] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.228s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.022835] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.025844] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.006s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.025844] env[63355]: DEBUG nova.objects.instance [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'resources' on Instance uuid 386e847e-967b-4247-9730-cdc5ac251474 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.059985] env[63355]: INFO nova.scheduler.client.report [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Deleted allocations for instance b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda [ 964.063396] env[63355]: INFO nova.scheduler.client.report [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Deleted allocations for instance 6706c65c-3959-471b-82c0-2196f4aa1ab0 [ 964.066369] env[63355]: DEBUG nova.compute.manager [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 964.238888] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349870, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.407466] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349874, 'name': PowerOffVM_Task, 'duration_secs': 0.236665} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.412075] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.413112] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6210bdc7-7a26-4225-a1f4-6c0363ae549b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.425219] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349875, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.441972] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ca2909-bcb2-48a7-99f8-b307677479fe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.483813] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.487015] env[63355]: DEBUG oslo_concurrency.lockutils [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.487015] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-289f8cfd-7172-4bd7-bf53-ca6da6fb5e2e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.500515] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 964.500515] env[63355]: value = "task-1349876" [ 964.500515] env[63355]: _type = "Task" [ 964.500515] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.514183] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] VM already powered off {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 964.514183] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 964.514317] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.514373] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.514676] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 964.514905] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0488b02-f477-4da3-aff3-82aa77d01e25 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.530732] env[63355]: DEBUG nova.compute.utils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 964.530732] env[63355]: DEBUG nova.compute.manager [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 964.530732] env[63355]: DEBUG nova.network.neutron [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 964.535069] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 964.535289] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 964.537123] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5938375-0eda-4d79-9750-f33eeb443755 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.549226] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 964.549226] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5243a085-e88c-9b28-ec4f-d1faf8eebebe" [ 964.549226] env[63355]: _type = "Task" [ 964.549226] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.564747] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5243a085-e88c-9b28-ec4f-d1faf8eebebe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.588311] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d1f9ac9-fd04-4f90-9d78-3a0637f1fe79 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.778s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.599733] env[63355]: DEBUG oslo_concurrency.lockutils [None req-27526a2a-9a02-40cb-9fa3-6413a13bd5dc tempest-ServersTestMultiNic-710959120 tempest-ServersTestMultiNic-710959120-project-member] Lock "6706c65c-3959-471b-82c0-2196f4aa1ab0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.952s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.621692] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.688862] env[63355]: DEBUG nova.policy [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d48712ba84642d1aab85b9976659a5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58d215a4ff55488f931814352915d256', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 964.742145] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349870, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.977184} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.742502] env[63355]: INFO nova.virt.vmwareapi.ds_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461/OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461.vmdk to [datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff/f755e516-c1d3-4618-aedf-9caa138936ff.vmdk. [ 964.743299] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Cleaning up location [datastore2] OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461 {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 964.743299] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_f61b1463-1e09-4be7-afcf-cc1ba6820461 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.743299] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2890be9b-3f20-4ce1-9df4-3a955ae558ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.760153] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 964.760153] env[63355]: value = "task-1349877" [ 964.760153] env[63355]: _type = "Task" [ 964.760153] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.772247] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349877, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.790370] env[63355]: DEBUG oslo_concurrency.lockutils [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "0b660647-697a-445b-bca4-82ef6975ea75" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.790370] env[63355]: DEBUG oslo_concurrency.lockutils [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "0b660647-697a-445b-bca4-82ef6975ea75" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.790370] env[63355]: DEBUG oslo_concurrency.lockutils [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "0b660647-697a-445b-bca4-82ef6975ea75-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.790370] env[63355]: DEBUG oslo_concurrency.lockutils [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "0b660647-697a-445b-bca4-82ef6975ea75-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.790370] env[63355]: DEBUG oslo_concurrency.lockutils [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "0b660647-697a-445b-bca4-82ef6975ea75-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.792746] env[63355]: INFO nova.compute.manager [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Terminating instance [ 964.796654] env[63355]: DEBUG nova.compute.manager [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 964.796929] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 964.797965] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729870c6-e3cb-4bf4-a36c-54b74ac8d857 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.818460] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.818460] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af5e3223-6e54-4ae8-9d2c-90cd4e441c4d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.830280] env[63355]: DEBUG oslo_vmware.api [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 964.830280] env[63355]: value = "task-1349878" [ 964.830280] env[63355]: _type = "Task" [ 964.830280] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.850106] env[63355]: DEBUG oslo_vmware.api [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349878, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.880505] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6a5ea6-a173-401d-a940-ec8cc1600efd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.888352] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7b4c0a-d776-40ee-bc1a-261e14a503eb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.929404] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d0923a-564b-4635-a944-7811251b3461 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.941515] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349875, 'name': ReconfigVM_Task, 'duration_secs': 0.964337} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.943377] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 1570cb36-76e0-4d06-8080-735b5246e92e/1570cb36-76e0-4d06-8080-735b5246e92e.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.944114] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55ea7ef-3117-47aa-8c18-6baa3a9169fc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.949529] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3c826ba-9450-41e7-9052-8f375888df69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.966308] env[63355]: DEBUG nova.compute.provider_tree [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 964.968494] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 964.968494] env[63355]: value = "task-1349879" [ 964.968494] env[63355]: _type = "Task" [ 964.968494] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.978284] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349879, 'name': Rename_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.037797] env[63355]: DEBUG nova.compute.manager [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 965.061090] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5243a085-e88c-9b28-ec4f-d1faf8eebebe, 'name': SearchDatastore_Task, 'duration_secs': 0.014854} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.061950] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c174175-158d-4a1d-badf-40eee4aabf8a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.070250] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 965.070250] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5216155f-60df-d6d6-8320-f993aa3b478b" [ 965.070250] env[63355]: _type = "Task" [ 965.070250] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.080615] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5216155f-60df-d6d6-8320-f993aa3b478b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.268899] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349877, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.04752} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.269288] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.269392] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff/f755e516-c1d3-4618-aedf-9caa138936ff.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.269667] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff/f755e516-c1d3-4618-aedf-9caa138936ff.vmdk to [datastore2] 248ec2f2-3a41-444a-b550-71c5438a45f7/248ec2f2-3a41-444a-b550-71c5438a45f7.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 965.269945] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1c5a58ad-15a7-40fd-a469-72351258a3d6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.283841] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 965.283841] env[63355]: value = "task-1349880" [ 965.283841] env[63355]: _type = "Task" [ 965.283841] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.309123] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349880, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.342637] env[63355]: DEBUG oslo_vmware.api [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349878, 'name': PowerOffVM_Task, 'duration_secs': 0.242079} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.342979] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.343189] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.343545] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25718d8f-eb3e-416c-8eda-11685a10d591 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.419379] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 965.419379] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 965.419379] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Deleting the datastore file [datastore1] 0b660647-697a-445b-bca4-82ef6975ea75 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.419744] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e0f2a13-c3ad-4f90-a366-4994591acf7e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.429649] env[63355]: DEBUG oslo_vmware.api [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for the task: (returnval){ [ 965.429649] env[63355]: value = "task-1349882" [ 965.429649] env[63355]: _type = "Task" [ 965.429649] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.440518] env[63355]: DEBUG oslo_vmware.api [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.491179] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349879, 'name': Rename_Task, 'duration_secs': 0.192076} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.491575] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.494278] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7821f87c-19c5-4109-b902-5a7488136e08 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.494278] env[63355]: DEBUG nova.compute.manager [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.495120] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451c7eab-7833-4ded-b5cd-cf2d35b433bb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.508808] env[63355]: ERROR nova.scheduler.client.report [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [req-a02cdd22-47c2-4a83-9cf6-03ccc5bd4ffd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 47abb610-db7e-4770-911d-187dd075ef8b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a02cdd22-47c2-4a83-9cf6-03ccc5bd4ffd"}]} [ 965.509472] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 965.509472] env[63355]: value = "task-1349883" [ 965.509472] env[63355]: _type = "Task" [ 965.509472] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.525532] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349883, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.532416] env[63355]: DEBUG nova.scheduler.client.report [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Refreshing inventories for resource provider 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 965.558147] env[63355]: DEBUG nova.scheduler.client.report [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Updating ProviderTree inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 965.558147] env[63355]: DEBUG nova.compute.provider_tree [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 965.578987] env[63355]: DEBUG nova.scheduler.client.report [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Refreshing aggregate associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, aggregates: None {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 965.585453] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5216155f-60df-d6d6-8320-f993aa3b478b, 'name': SearchDatastore_Task, 'duration_secs': 0.01169} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.585453] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.585795] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 8e9ff9b7-636f-48df-9168-509d733278f8/84ca0bb3-9916-4cea-9399-75af77b0558e-rescue.vmdk. {{(pid=63355) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 965.586028] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d24d69c-3277-46ba-8fce-c858c8f94afa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.598124] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 965.598124] env[63355]: value = "task-1349884" [ 965.598124] env[63355]: _type = "Task" [ 965.598124] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.605013] env[63355]: DEBUG nova.scheduler.client.report [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Refreshing trait associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 965.614525] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.799879] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349880, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.884472] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904406c5-f608-4ce9-b5f1-0c6faa739289 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.888588] env[63355]: DEBUG nova.network.neutron [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Successfully created port: d4c4170a-d27a-4f64-8789-6e01e1adb3df {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.900164] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bff9756-2867-4a07-859e-c171fe42f449 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.951798] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0c698f-1166-48ed-bf54-fa5a46fc078d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.963586] env[63355]: DEBUG oslo_vmware.api [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Task: {'id': task-1349882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227193} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.966334] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.966616] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.966807] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.967022] env[63355]: INFO nova.compute.manager [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Took 1.17 seconds to destroy the instance on the hypervisor. [ 965.967304] env[63355]: DEBUG oslo.service.loopingcall [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.967626] env[63355]: DEBUG nova.compute.manager [-] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.967720] env[63355]: DEBUG nova.network.neutron [-] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 965.971713] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52347ca-28bb-493e-825d-9761549c3023 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.990573] env[63355]: DEBUG nova.compute.provider_tree [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 966.013645] env[63355]: INFO nova.compute.manager [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] instance snapshotting [ 966.014517] env[63355]: DEBUG nova.objects.instance [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'flavor' on Instance uuid 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.032409] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349883, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.055165] env[63355]: DEBUG nova.compute.manager [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.090040] env[63355]: DEBUG nova.virt.hardware [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.091392] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3233bdbf-3872-4b6e-b7e8-1f9ff970faa0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.106785] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fbc773-0afa-406c-9d49-7018a8ea39ee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.124330] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349884, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.308682] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349880, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.523465] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b00bbe-ba65-4eec-b82d-0a5d64c06f46 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.532090] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349883, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.546664] env[63355]: DEBUG nova.scheduler.client.report [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Updated inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b with generation 97 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 966.546967] env[63355]: DEBUG nova.compute.provider_tree [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Updating resource provider 47abb610-db7e-4770-911d-187dd075ef8b generation from 97 to 98 during operation: update_inventory {{(pid=63355) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 966.547138] env[63355]: DEBUG nova.compute.provider_tree [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 966.551428] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6e9d00-56c4-4e71-aa30-93cce43c73a0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.612807] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349884, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.711207} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.613157] env[63355]: INFO nova.virt.vmwareapi.ds_util [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 8e9ff9b7-636f-48df-9168-509d733278f8/84ca0bb3-9916-4cea-9399-75af77b0558e-rescue.vmdk. [ 966.617020] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1ef131-000f-4b3f-99bb-be7ebaefda51 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.644215] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 8e9ff9b7-636f-48df-9168-509d733278f8/84ca0bb3-9916-4cea-9399-75af77b0558e-rescue.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.644677] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34c7a1be-d219-482d-845d-adeb26891a78 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.674758] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 966.674758] env[63355]: value = "task-1349885" [ 966.674758] env[63355]: _type = "Task" [ 966.674758] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.684457] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349885, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.799055] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349880, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.969198] env[63355]: DEBUG nova.compute.manager [req-0b7a1c0d-2465-48c9-80f1-59ade4619094 req-418232c7-6d0f-46ed-b7fa-ceae6ecb1a6c service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Received event network-vif-deleted-86bfab04-fd0a-4339-922e-241bb081e78e {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.969455] env[63355]: INFO nova.compute.manager [req-0b7a1c0d-2465-48c9-80f1-59ade4619094 req-418232c7-6d0f-46ed-b7fa-ceae6ecb1a6c service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Neutron deleted interface 86bfab04-fd0a-4339-922e-241bb081e78e; detaching it from the instance and deleting it from the info cache [ 966.969632] env[63355]: DEBUG nova.network.neutron [req-0b7a1c0d-2465-48c9-80f1-59ade4619094 req-418232c7-6d0f-46ed-b7fa-ceae6ecb1a6c service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.030158] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349883, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.056047] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.030s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.059955] env[63355]: DEBUG oslo_concurrency.lockutils [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.575s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.059955] env[63355]: DEBUG nova.objects.instance [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lazy-loading 'resources' on Instance uuid fc6a99e6-5319-47d3-8175-770c06d3e325 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.069364] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 967.070438] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e6a8848c-f062-4860-9f0b-e523e632bfe0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.082243] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 967.082243] env[63355]: value = "task-1349886" [ 967.082243] env[63355]: _type = "Task" [ 967.082243] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.096552] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349886, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.110106] env[63355]: INFO nova.scheduler.client.report [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Deleted allocations for instance 386e847e-967b-4247-9730-cdc5ac251474 [ 967.187945] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349885, 'name': ReconfigVM_Task, 'duration_secs': 0.455179} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.188463] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 8e9ff9b7-636f-48df-9168-509d733278f8/84ca0bb3-9916-4cea-9399-75af77b0558e-rescue.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.189386] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f86ea5-03e3-4364-9bdf-420b87d96bde {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.222399] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e34eea9-7605-4a35-a5c0-f8ca1f075691 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.242303] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 967.242303] env[63355]: value = "task-1349887" [ 967.242303] env[63355]: _type = "Task" [ 967.242303] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.256618] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349887, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.306964] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349880, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.362160] env[63355]: DEBUG nova.network.neutron [-] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.471973] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c321b90-6c2b-4838-b47a-6a439bc56877 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.485683] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7948830-af37-4cbb-a63f-02dd647b6e9a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.527985] env[63355]: DEBUG nova.compute.manager [req-0b7a1c0d-2465-48c9-80f1-59ade4619094 req-418232c7-6d0f-46ed-b7fa-ceae6ecb1a6c service nova] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Detach interface failed, port_id=86bfab04-fd0a-4339-922e-241bb081e78e, reason: Instance 0b660647-697a-445b-bca4-82ef6975ea75 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 967.538887] env[63355]: DEBUG oslo_vmware.api [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349883, 'name': PowerOnVM_Task, 'duration_secs': 1.5426} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.539206] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 967.539416] env[63355]: DEBUG nova.compute.manager [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.541495] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209362e4-3e24-4604-ab0d-87652c135b57 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.603074] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349886, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.620229] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fc15fde7-1752-4548-88a7-327acfe6e07a tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "386e847e-967b-4247-9730-cdc5ac251474" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.623s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.765317] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349887, 'name': ReconfigVM_Task, 'duration_secs': 0.388033} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.765817] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 967.766288] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57074c5c-aece-4617-ad64-2d18155dcd73 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.776128] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 967.776128] env[63355]: value = "task-1349888" [ 967.776128] env[63355]: _type = "Task" [ 967.776128] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.794543] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349888, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.804740] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349880, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.822374] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d80ba9-0524-4b9f-ab62-b7fcf68a9367 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.830160] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7bac149-2eeb-437a-8353-d77ca47a6be1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.864867] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774d8389-8b95-4a7d-8d7d-20d297966f5c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.868334] env[63355]: INFO nova.compute.manager [-] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Took 1.90 seconds to deallocate network for instance. [ 967.879832] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9ac230-7dfd-48cc-af7e-b7f259521c08 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.900665] env[63355]: DEBUG nova.compute.provider_tree [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.059338] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.099464] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349886, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.294866] env[63355]: DEBUG nova.network.neutron [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Successfully updated port: d4c4170a-d27a-4f64-8789-6e01e1adb3df {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 968.314536] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349888, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.314839] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349880, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.633752} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.315780] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f755e516-c1d3-4618-aedf-9caa138936ff/f755e516-c1d3-4618-aedf-9caa138936ff.vmdk to [datastore2] 248ec2f2-3a41-444a-b550-71c5438a45f7/248ec2f2-3a41-444a-b550-71c5438a45f7.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.316658] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eaeb14d-bbd0-4a75-adae-9d993ea5e9a3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.346211] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 248ec2f2-3a41-444a-b550-71c5438a45f7/248ec2f2-3a41-444a-b550-71c5438a45f7.vmdk or device None with type streamOptimized {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.346898] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-423090db-5643-4e9a-be0b-c8d3495fb9ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.368470] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 968.368470] env[63355]: value = "task-1349889" [ 968.368470] env[63355]: _type = "Task" [ 968.368470] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.379351] env[63355]: DEBUG oslo_concurrency.lockutils [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.379703] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349889, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.404413] env[63355]: DEBUG nova.scheduler.client.report [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.598312] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349886, 'name': CreateSnapshot_Task, 'duration_secs': 1.119708} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.598907] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 968.600136] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1893e37-a08d-4973-ad0d-3e41284af5e8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.789591] env[63355]: DEBUG oslo_vmware.api [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349888, 'name': PowerOnVM_Task, 'duration_secs': 0.927015} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.789950] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 968.794270] env[63355]: DEBUG nova.compute.manager [None req-7c88dbcf-654f-4f7f-9525-8a978e90f9dc tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.795573] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd71129-a0aa-40c0-849a-96710d05b50c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.800467] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.800467] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquired lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.800467] env[63355]: DEBUG nova.network.neutron [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.887297] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349889, 'name': ReconfigVM_Task, 'duration_secs': 0.383723} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.887721] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 248ec2f2-3a41-444a-b550-71c5438a45f7/248ec2f2-3a41-444a-b550-71c5438a45f7.vmdk or device None with type streamOptimized {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 968.888544] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-152942ec-9824-466c-9e8d-88fcf74e6b7f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.896774] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 968.896774] env[63355]: value = "task-1349890" [ 968.896774] env[63355]: _type = "Task" [ 968.896774] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.911376] env[63355]: DEBUG oslo_concurrency.lockutils [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.913494] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349890, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.914481] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.293s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.920572] env[63355]: INFO nova.compute.claims [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 968.971575] env[63355]: INFO nova.scheduler.client.report [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted allocations for instance fc6a99e6-5319-47d3-8175-770c06d3e325 [ 969.033537] env[63355]: DEBUG nova.compute.manager [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Received event network-vif-plugged-d4c4170a-d27a-4f64-8789-6e01e1adb3df {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.033537] env[63355]: DEBUG oslo_concurrency.lockutils [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] Acquiring lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.033913] env[63355]: DEBUG oslo_concurrency.lockutils [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.033913] env[63355]: DEBUG oslo_concurrency.lockutils [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.034026] env[63355]: DEBUG nova.compute.manager [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] No waiting events found dispatching network-vif-plugged-d4c4170a-d27a-4f64-8789-6e01e1adb3df {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 969.034248] env[63355]: WARNING nova.compute.manager [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Received unexpected event network-vif-plugged-d4c4170a-d27a-4f64-8789-6e01e1adb3df for instance with vm_state building and task_state spawning. [ 969.034473] env[63355]: DEBUG nova.compute.manager [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Received event network-changed-d4c4170a-d27a-4f64-8789-6e01e1adb3df {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.034567] env[63355]: DEBUG nova.compute.manager [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Refreshing instance network info cache due to event network-changed-d4c4170a-d27a-4f64-8789-6e01e1adb3df. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 969.034924] env[63355]: DEBUG oslo_concurrency.lockutils [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] Acquiring lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.125364] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 969.127329] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-378a5f98-69b0-47fc-b14d-712102fddfdf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.139706] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 969.139706] env[63355]: value = "task-1349891" [ 969.139706] env[63355]: _type = "Task" [ 969.139706] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.152554] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349891, 'name': CloneVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.387300] env[63355]: DEBUG nova.network.neutron [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 969.414506] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349890, 'name': Rename_Task, 'duration_secs': 0.172167} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.414916] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.415318] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2f02a81-b895-42e5-8813-e375338b38e0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.423541] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 969.423541] env[63355]: value = "task-1349892" [ 969.423541] env[63355]: _type = "Task" [ 969.423541] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.441775] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349892, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.486034] env[63355]: DEBUG oslo_concurrency.lockutils [None req-71e59117-9015-430b-83a8-3248f6df2ecd tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "fc6a99e6-5319-47d3-8175-770c06d3e325" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.353s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.584223] env[63355]: DEBUG nova.network.neutron [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updating instance_info_cache with network_info: [{"id": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "address": "fa:16:3e:fb:af:dc", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4c4170a-d2", "ovs_interfaceid": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.604333] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "1570cb36-76e0-4d06-8080-735b5246e92e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.604620] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "1570cb36-76e0-4d06-8080-735b5246e92e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.604844] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "1570cb36-76e0-4d06-8080-735b5246e92e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.605054] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "1570cb36-76e0-4d06-8080-735b5246e92e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.605360] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "1570cb36-76e0-4d06-8080-735b5246e92e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.612834] env[63355]: INFO nova.compute.manager [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Terminating instance [ 969.613578] env[63355]: DEBUG nova.compute.manager [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.613845] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.614691] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d61224-931b-479c-bbf7-d199a553d82d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.626551] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.626551] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33d80151-6feb-4a61-8c08-0e30796d6c38 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.636937] env[63355]: DEBUG oslo_vmware.api [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 969.636937] env[63355]: value = "task-1349893" [ 969.636937] env[63355]: _type = "Task" [ 969.636937] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.651953] env[63355]: DEBUG oslo_vmware.api [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.655937] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349891, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.944051] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349892, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.090886] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Releasing lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.091050] env[63355]: DEBUG nova.compute.manager [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Instance network_info: |[{"id": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "address": "fa:16:3e:fb:af:dc", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4c4170a-d2", "ovs_interfaceid": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 970.091349] env[63355]: DEBUG oslo_concurrency.lockutils [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] Acquired lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.091558] env[63355]: DEBUG nova.network.neutron [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Refreshing network info cache for port d4c4170a-d27a-4f64-8789-6e01e1adb3df {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 970.092901] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:af:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd986680e-ad16-45b1-bf6d-cd2fe661679f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4c4170a-d27a-4f64-8789-6e01e1adb3df', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 970.102134] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Creating folder: Project (58d215a4ff55488f931814352915d256). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 970.105469] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a1ab1d28-823a-4e1e-872b-7505eee20cb8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.125150] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Created folder: Project (58d215a4ff55488f931814352915d256) in parent group-v287607. [ 970.127698] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Creating folder: Instances. Parent ref: group-v287716. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 970.127698] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-acc35180-9f2c-45f6-9134-fa5ddd98e844 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.137074] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Created folder: Instances in parent group-v287716. [ 970.137380] env[63355]: DEBUG oslo.service.loopingcall [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.140674] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 970.144647] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65cd9ac7-2b5c-43ed-b723-45311826fd19 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.163992] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df227e0e-9d39-4ad0-92f5-85853ba35eab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.181823] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9984bbdc-0cf5-4f0c-a0d3-e2073f844d03 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.185185] env[63355]: DEBUG oslo_vmware.api [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349893, 'name': PowerOffVM_Task, 'duration_secs': 0.238878} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.186996] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.187238] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.187490] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 970.187490] env[63355]: value = "task-1349896" [ 970.187490] env[63355]: _type = "Task" [ 970.187490] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.188064] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09a84087-59d5-4a7a-9904-1b5e0b0a65e5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.196919] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349891, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.225147] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456b73d7-2417-456b-9c05-ef0e5c15808e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.231232] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349896, 'name': CreateVM_Task} progress is 15%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.240208] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c1eba5-1733-47b7-a8ac-af8ca2b4d84a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.259595] env[63355]: DEBUG nova.compute.provider_tree [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.288218] env[63355]: INFO nova.compute.manager [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Unrescuing [ 970.288603] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.288792] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquired lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.288994] env[63355]: DEBUG nova.network.neutron [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.313390] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.313708] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.313829] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleting the datastore file [datastore2] 1570cb36-76e0-4d06-8080-735b5246e92e {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.314458] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af3c8ffc-98b2-4331-84ae-caf0d6dc069d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.322682] env[63355]: DEBUG oslo_vmware.api [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 970.322682] env[63355]: value = "task-1349898" [ 970.322682] env[63355]: _type = "Task" [ 970.322682] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.332306] env[63355]: DEBUG oslo_vmware.api [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.439730] env[63355]: DEBUG oslo_vmware.api [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349892, 'name': PowerOnVM_Task, 'duration_secs': 0.542196} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.439730] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.439730] env[63355]: INFO nova.compute.manager [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Took 15.49 seconds to spawn the instance on the hypervisor. [ 970.439730] env[63355]: DEBUG nova.compute.manager [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.440552] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93bfc5f7-c0ac-4688-8398-757706e970fe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.505669] env[63355]: DEBUG nova.network.neutron [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updated VIF entry in instance network info cache for port d4c4170a-d27a-4f64-8789-6e01e1adb3df. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.506222] env[63355]: DEBUG nova.network.neutron [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updating instance_info_cache with network_info: [{"id": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "address": "fa:16:3e:fb:af:dc", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4c4170a-d2", "ovs_interfaceid": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.675963] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349891, 'name': CloneVM_Task, 'duration_secs': 1.510803} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.676333] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Created linked-clone VM from snapshot [ 970.677451] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2bc4e8-6f22-486f-b088-e1ecb487a2f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.687470] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Uploading image eabfdc0f-53a1-4aa1-a863-de4d0008a5e5 {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 970.701533] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349896, 'name': CreateVM_Task, 'duration_secs': 0.455568} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.701774] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 970.702981] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.703295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.703652] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 970.703932] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06c26d85-4d34-4c1c-a916-a335a23eaa63 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.712306] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 970.712306] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d9743f-d437-a28a-5f23-f806663e7335" [ 970.712306] env[63355]: _type = "Task" [ 970.712306] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.729143] env[63355]: DEBUG oslo_vmware.rw_handles [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 970.729143] env[63355]: value = "vm-287715" [ 970.729143] env[63355]: _type = "VirtualMachine" [ 970.729143] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 970.733518] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-135cf2e6-e47f-4a0a-b4d9-1664f181ac42 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.735282] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d9743f-d437-a28a-5f23-f806663e7335, 'name': SearchDatastore_Task, 'duration_secs': 0.011594} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.735592] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.735857] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.736134] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.736303] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.736492] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.737134] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dbf3a0f0-6082-474b-9629-6334be60e061 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.744498] env[63355]: DEBUG oslo_vmware.rw_handles [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lease: (returnval){ [ 970.744498] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524728aa-6d31-fc2c-0aa4-9ecf3f0b48df" [ 970.744498] env[63355]: _type = "HttpNfcLease" [ 970.744498] env[63355]: } obtained for exporting VM: (result){ [ 970.744498] env[63355]: value = "vm-287715" [ 970.744498] env[63355]: _type = "VirtualMachine" [ 970.744498] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 970.744860] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the lease: (returnval){ [ 970.744860] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524728aa-6d31-fc2c-0aa4-9ecf3f0b48df" [ 970.744860] env[63355]: _type = "HttpNfcLease" [ 970.744860] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 970.750658] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.750898] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 970.752020] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a926c758-ab80-4066-9e71-d5a5b535f510 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.756532] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 970.756532] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524728aa-6d31-fc2c-0aa4-9ecf3f0b48df" [ 970.756532] env[63355]: _type = "HttpNfcLease" [ 970.756532] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 970.759613] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 970.759613] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52935c29-8569-4471-78b3-ff01443016c5" [ 970.759613] env[63355]: _type = "Task" [ 970.759613] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.764533] env[63355]: DEBUG nova.scheduler.client.report [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.777242] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52935c29-8569-4471-78b3-ff01443016c5, 'name': SearchDatastore_Task, 'duration_secs': 0.011029} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.777242] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-489b2a84-7f14-4d7a-a70e-f8d1796728f6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.782242] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 970.782242] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5252ff7d-47f6-d22f-0f1c-fcb96245ef9f" [ 970.782242] env[63355]: _type = "Task" [ 970.782242] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.797225] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5252ff7d-47f6-d22f-0f1c-fcb96245ef9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.833933] env[63355]: DEBUG oslo_vmware.api [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234721} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.834574] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.834813] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.835015] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.835211] env[63355]: INFO nova.compute.manager [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Took 1.22 seconds to destroy the instance on the hypervisor. [ 970.835481] env[63355]: DEBUG oslo.service.loopingcall [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.835662] env[63355]: DEBUG nova.compute.manager [-] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.835757] env[63355]: DEBUG nova.network.neutron [-] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 970.977448] env[63355]: INFO nova.compute.manager [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Took 39.04 seconds to build instance. [ 971.017381] env[63355]: DEBUG oslo_concurrency.lockutils [req-6462b94e-cb77-463c-b50c-af2deaf22347 req-35c034bb-8039-4e78-a128-2aaaa85430ed service nova] Releasing lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.026673] env[63355]: DEBUG oslo_vmware.rw_handles [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52278bd2-2895-8cc5-e798-a1b7f9aad49c/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 971.028354] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26382bc4-733c-4ffa-9447-b66063b605f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.038994] env[63355]: DEBUG oslo_vmware.rw_handles [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52278bd2-2895-8cc5-e798-a1b7f9aad49c/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 971.038994] env[63355]: ERROR oslo_vmware.rw_handles [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52278bd2-2895-8cc5-e798-a1b7f9aad49c/disk-0.vmdk due to incomplete transfer. [ 971.038994] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-df80812e-2be5-4035-9293-18160d98cc7f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.053714] env[63355]: DEBUG oslo_vmware.rw_handles [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52278bd2-2895-8cc5-e798-a1b7f9aad49c/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 971.053873] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Uploaded image 95245748-2b88-4a4c-8d0b-6a828a676666 to the Glance image server {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 971.059333] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 971.059333] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a7ef8ad1-98de-4063-91a7-426083edfd67 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.068012] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 971.068012] env[63355]: value = "task-1349900" [ 971.068012] env[63355]: _type = "Task" [ 971.068012] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.087411] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349900, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.243796] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "fd55ddfc-f376-4e15-961e-6826f31a7890" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.246213] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.260663] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 971.260663] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524728aa-6d31-fc2c-0aa4-9ecf3f0b48df" [ 971.260663] env[63355]: _type = "HttpNfcLease" [ 971.260663] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 971.261734] env[63355]: DEBUG oslo_vmware.rw_handles [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 971.261734] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524728aa-6d31-fc2c-0aa4-9ecf3f0b48df" [ 971.261734] env[63355]: _type = "HttpNfcLease" [ 971.261734] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 971.262806] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964981b7-04e9-47b1-8e6e-27b3dc080362 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.270354] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.270664] env[63355]: DEBUG nova.compute.manager [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 971.279662] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 3.218s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.279662] env[63355]: DEBUG nova.objects.instance [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63355) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 971.280317] env[63355]: DEBUG oslo_vmware.rw_handles [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52205b72-7bf7-a082-bdae-de171644ed4c/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 971.280500] env[63355]: DEBUG oslo_vmware.rw_handles [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52205b72-7bf7-a082-bdae-de171644ed4c/disk-0.vmdk for reading. {{(pid=63355) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 971.361209] env[63355]: DEBUG nova.network.neutron [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Updating instance_info_cache with network_info: [{"id": "a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a", "address": "fa:16:3e:f8:de:e7", "network": {"id": "26b0dabd-e662-4b47-bcd4-a50e05f0a52d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-825312565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1efb1730acbf4ba5a19bd150bf99d48f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0ceb2ca-35", "ovs_interfaceid": "a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.371737] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5252ff7d-47f6-d22f-0f1c-fcb96245ef9f, 'name': SearchDatastore_Task, 'duration_secs': 0.011373} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.372011] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.372284] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] ae3961d2-dc5b-4e49-acca-6fb52291f23b/ae3961d2-dc5b-4e49-acca-6fb52291f23b.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 971.372554] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-006e4db8-1f8e-4d42-8d34-3ea729ec8475 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.381752] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 971.381752] env[63355]: value = "task-1349901" [ 971.381752] env[63355]: _type = "Task" [ 971.381752] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.392260] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349901, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.433181] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6541b264-9cf2-49ed-8b7b-fab23b4c0aaa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.480586] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ea2fde47-eae7-414a-b0ff-9a502dfc14dc tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "248ec2f2-3a41-444a-b550-71c5438a45f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.827s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.581103] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349900, 'name': Destroy_Task, 'duration_secs': 0.505347} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.581103] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Destroyed the VM [ 971.581426] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 971.581908] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-541f6320-353e-4b79-9dd6-a8979288df19 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.592811] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 971.592811] env[63355]: value = "task-1349902" [ 971.592811] env[63355]: _type = "Task" [ 971.592811] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.603907] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349902, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.755859] env[63355]: DEBUG nova.compute.manager [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.787431] env[63355]: DEBUG nova.compute.utils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 971.793944] env[63355]: DEBUG nova.compute.manager [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 971.793944] env[63355]: DEBUG nova.network.neutron [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 971.818408] env[63355]: DEBUG nova.compute.manager [req-a0775bd6-4c03-4756-86fe-7ec86156ea05 req-adebfdad-56b6-46cd-acb3-074b2aa02fb4 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Received event network-vif-deleted-91c7c439-7d4c-4512-90c6-d1698630ae2b {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.818831] env[63355]: INFO nova.compute.manager [req-a0775bd6-4c03-4756-86fe-7ec86156ea05 req-adebfdad-56b6-46cd-acb3-074b2aa02fb4 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Neutron deleted interface 91c7c439-7d4c-4512-90c6-d1698630ae2b; detaching it from the instance and deleting it from the info cache [ 971.818976] env[63355]: DEBUG nova.network.neutron [req-a0775bd6-4c03-4756-86fe-7ec86156ea05 req-adebfdad-56b6-46cd-acb3-074b2aa02fb4 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.850713] env[63355]: DEBUG nova.policy [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a40a28e383ff477abd2a21ade8c71012', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '226038844bda45b28033d6ea40a9bc9b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 971.862217] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "edb14fe7-d444-4fef-8c5d-d5616676e326" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.862707] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "edb14fe7-d444-4fef-8c5d-d5616676e326" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.865809] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Releasing lock "refresh_cache-8e9ff9b7-636f-48df-9168-509d733278f8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.868446] env[63355]: DEBUG nova.objects.instance [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lazy-loading 'flavor' on Instance uuid 8e9ff9b7-636f-48df-9168-509d733278f8 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.898698] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349901, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506598} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.899179] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] ae3961d2-dc5b-4e49-acca-6fb52291f23b/ae3961d2-dc5b-4e49-acca-6fb52291f23b.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 971.899617] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 971.899976] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45e1d94f-2881-4773-9e7b-4a4d898f22a9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.911223] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 971.911223] env[63355]: value = "task-1349903" [ 971.911223] env[63355]: _type = "Task" [ 971.911223] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.925531] env[63355]: DEBUG nova.network.neutron [-] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.927717] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349903, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.107041] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349902, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.276191] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "248ec2f2-3a41-444a-b550-71c5438a45f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.276191] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "248ec2f2-3a41-444a-b550-71c5438a45f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.276191] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "248ec2f2-3a41-444a-b550-71c5438a45f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.276761] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "248ec2f2-3a41-444a-b550-71c5438a45f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.277079] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "248ec2f2-3a41-444a-b550-71c5438a45f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.280348] env[63355]: INFO nova.compute.manager [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Terminating instance [ 972.284137] env[63355]: DEBUG nova.compute.manager [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 972.284191] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 972.285928] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.287290] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e2cb76-b894-4fe5-b919-77214eb31a08 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.292263] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b19c719b-e304-402c-b80d-83b8a306d05c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.296492] env[63355]: DEBUG oslo_concurrency.lockutils [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.917s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.296855] env[63355]: DEBUG nova.objects.instance [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lazy-loading 'resources' on Instance uuid 0b660647-697a-445b-bca4-82ef6975ea75 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.299957] env[63355]: DEBUG nova.compute.manager [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 972.304586] env[63355]: DEBUG nova.network.neutron [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Successfully created port: eb47f210-8e1e-4310-b314-33a489cf4325 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 972.309266] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 972.310758] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c727a6c-d85a-4b0d-a85c-7adbf4b57df7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.319400] env[63355]: DEBUG oslo_vmware.api [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 972.319400] env[63355]: value = "task-1349904" [ 972.319400] env[63355]: _type = "Task" [ 972.319400] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.324061] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-afb7527d-64c4-4641-b288-854062d32100 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.332524] env[63355]: DEBUG oslo_vmware.api [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349904, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.337669] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf8dfba-9a44-49c6-ad40-91a8e06a34da {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.373934] env[63355]: DEBUG nova.compute.manager [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.381090] env[63355]: DEBUG nova.compute.manager [req-a0775bd6-4c03-4756-86fe-7ec86156ea05 req-adebfdad-56b6-46cd-acb3-074b2aa02fb4 service nova] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Detach interface failed, port_id=91c7c439-7d4c-4512-90c6-d1698630ae2b, reason: Instance 1570cb36-76e0-4d06-8080-735b5246e92e could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 972.382197] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd75e96-6b24-4e39-b452-9708069cb093 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.406440] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 972.406440] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b40089f-81d3-4764-8405-f051e48984b6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.416273] env[63355]: DEBUG oslo_vmware.api [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 972.416273] env[63355]: value = "task-1349905" [ 972.416273] env[63355]: _type = "Task" [ 972.416273] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.427047] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349903, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072973} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.427702] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.428993] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc265f6-84a4-409b-8677-93dbbb7cfd77 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.435460] env[63355]: INFO nova.compute.manager [-] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Took 1.60 seconds to deallocate network for instance. [ 972.436265] env[63355]: DEBUG oslo_vmware.api [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349905, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.463015] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] ae3961d2-dc5b-4e49-acca-6fb52291f23b/ae3961d2-dc5b-4e49-acca-6fb52291f23b.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.463542] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95a468f4-6263-4192-9177-47418330731b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.484516] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 972.484516] env[63355]: value = "task-1349906" [ 972.484516] env[63355]: _type = "Task" [ 972.484516] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.492801] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349906, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.604384] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349902, 'name': RemoveSnapshot_Task, 'duration_secs': 0.813709} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.604979] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 972.604979] env[63355]: DEBUG nova.compute.manager [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.606102] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434fa379-4ad8-4b71-b770-1db11019ecd4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.677586] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.677910] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.745889] env[63355]: INFO nova.compute.manager [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Rebuilding instance [ 972.815042] env[63355]: DEBUG nova.compute.manager [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.816469] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-badd5960-caa5-4dd1-ad20-5b3271b0b81d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.834263] env[63355]: DEBUG oslo_vmware.api [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349904, 'name': PowerOffVM_Task, 'duration_secs': 0.193049} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.835204] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 972.835429] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 972.842219] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3cbb889-96f0-4c5b-ab28-0478028011f7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.898709] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.927641] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 972.927641] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 972.927641] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleting the datastore file [datastore2] 248ec2f2-3a41-444a-b550-71c5438a45f7 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.927641] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2c073f0-e0c0-47b0-b09e-0672c820c817 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.940825] env[63355]: DEBUG oslo_vmware.api [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349905, 'name': PowerOffVM_Task, 'duration_secs': 0.381951} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.952038] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.952180] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 972.958210] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Reconfiguring VM instance instance-00000051 to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 972.959118] env[63355]: DEBUG oslo_vmware.api [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 972.959118] env[63355]: value = "task-1349908" [ 972.959118] env[63355]: _type = "Task" [ 972.959118] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.960289] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d064f848-5b74-4e35-8046-f73492e0c8b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.991707] env[63355]: DEBUG oslo_vmware.api [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 972.991707] env[63355]: value = "task-1349909" [ 972.991707] env[63355]: _type = "Task" [ 972.991707] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.999693] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349906, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.999693] env[63355]: DEBUG oslo_vmware.api [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349908, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.009138] env[63355]: DEBUG oslo_vmware.api [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349909, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.119829] env[63355]: INFO nova.compute.manager [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Shelve offloading [ 973.122277] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.122686] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8445ea5-f084-446a-85cc-15bf2b9b5d63 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.130440] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4889f1b9-b449-4d92-933f-fe6f26c2e619 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.137027] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 973.137027] env[63355]: value = "task-1349910" [ 973.137027] env[63355]: _type = "Task" [ 973.137027] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.143726] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc94f8e-d0a1-4afd-b893-74e2545386de {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.155986] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] VM already powered off {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 973.155986] env[63355]: DEBUG nova.compute.manager [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.155986] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cfd6deb-e0e1-462f-b369-a8933d9d392c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.190361] env[63355]: DEBUG nova.compute.manager [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 973.197127] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8599bf-33b0-4bcd-a6e3-6a96a315d676 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.202703] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.202703] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.202703] env[63355]: DEBUG nova.network.neutron [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 973.209265] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541370ac-baef-457f-8590-da8ffb46f285 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.227115] env[63355]: DEBUG nova.compute.provider_tree [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.313953] env[63355]: DEBUG nova.compute.manager [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 973.337295] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 973.337791] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 973.337961] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 973.338166] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 973.338326] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 973.338480] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 973.338731] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 973.338929] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 973.339391] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 973.339608] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 973.339981] env[63355]: DEBUG nova.virt.hardware [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 973.341697] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.342704] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa6b334-5fde-4d94-93bf-fec70bd80769 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.345816] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5aa1991e-b5ce-4b8f-8ce1-2a220497a407 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.358852] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for the task: (returnval){ [ 973.358852] env[63355]: value = "task-1349911" [ 973.358852] env[63355]: _type = "Task" [ 973.358852] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.359035] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76030ea-dda0-479f-ab57-0aad7b1fb061 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.386135] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349911, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.488689] env[63355]: DEBUG oslo_vmware.api [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349908, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187397} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.492774] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.492999] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 973.493237] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 973.493460] env[63355]: INFO nova.compute.manager [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Took 1.21 seconds to destroy the instance on the hypervisor. [ 973.493843] env[63355]: DEBUG oslo.service.loopingcall [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.494035] env[63355]: DEBUG nova.compute.manager [-] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.494035] env[63355]: DEBUG nova.network.neutron [-] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 973.501590] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349906, 'name': ReconfigVM_Task, 'duration_secs': 0.775084} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.505820] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Reconfigured VM instance instance-00000053 to attach disk [datastore2] ae3961d2-dc5b-4e49-acca-6fb52291f23b/ae3961d2-dc5b-4e49-acca-6fb52291f23b.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.506640] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc5ce571-fcf3-42df-9ac4-f985e8f2ec30 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.515159] env[63355]: DEBUG oslo_vmware.api [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349909, 'name': ReconfigVM_Task, 'duration_secs': 0.438355} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.516787] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Reconfigured VM instance instance-00000051 to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 973.517098] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 973.517439] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 973.517439] env[63355]: value = "task-1349912" [ 973.517439] env[63355]: _type = "Task" [ 973.517439] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.517640] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d01f37e-dd04-493f-8ce9-9b325349084b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.530051] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349912, 'name': Rename_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.531377] env[63355]: DEBUG oslo_vmware.api [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 973.531377] env[63355]: value = "task-1349913" [ 973.531377] env[63355]: _type = "Task" [ 973.531377] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.542908] env[63355]: DEBUG oslo_vmware.api [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349913, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.713832] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.730638] env[63355]: DEBUG nova.scheduler.client.report [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.893971] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349911, 'name': PowerOffVM_Task, 'duration_secs': 0.208847} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.894436] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.895361] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.895669] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7cc9e9b6-b59e-4272-8120-0017a621595c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.907984] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for the task: (returnval){ [ 973.907984] env[63355]: value = "task-1349914" [ 973.907984] env[63355]: _type = "Task" [ 973.907984] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.920175] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] VM already powered off {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 973.920436] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 973.920739] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287631', 'volume_id': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'name': 'volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ecbacdb3-516e-43c3-96e3-4961b76565ca', 'attached_at': '', 'detached_at': '', 'volume_id': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'serial': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 973.921774] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af38ce9a-6dd5-4493-b7e7-d1d1d31a1d27 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.946947] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931ca49f-315f-40f6-88ab-59e984a2179c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.955741] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e439940-f76d-4759-8945-7ae1eded58ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.978048] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18aa870e-11e8-4ea0-b441-b16722498fbf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.993609] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] The volume has not been displaced from its original location: [datastore1] volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1/volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 973.999271] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Reconfiguring VM instance instance-00000049 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 973.999638] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99c69395-9ebd-4578-b5f3-9914c4a259e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.021741] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for the task: (returnval){ [ 974.021741] env[63355]: value = "task-1349915" [ 974.021741] env[63355]: _type = "Task" [ 974.021741] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.035419] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349912, 'name': Rename_Task, 'duration_secs': 0.15135} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.042534] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.042902] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349915, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.043482] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a70c2e85-0295-402c-a4f6-a508d1af139b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.051361] env[63355]: DEBUG oslo_vmware.api [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349913, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.052925] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 974.052925] env[63355]: value = "task-1349916" [ 974.052925] env[63355]: _type = "Task" [ 974.052925] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.061721] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349916, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.078678] env[63355]: DEBUG nova.network.neutron [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Updating instance_info_cache with network_info: [{"id": "850ef0bb-ad5f-4627-b482-a78817b84343", "address": "fa:16:3e:42:ee:f7", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap850ef0bb-ad", "ovs_interfaceid": "850ef0bb-ad5f-4627-b482-a78817b84343", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.237082] env[63355]: DEBUG oslo_concurrency.lockutils [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.940s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.241263] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.954s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.242103] env[63355]: INFO nova.compute.claims [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.274849] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquiring lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.275366] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.283238] env[63355]: DEBUG nova.compute.manager [req-283d51b8-80a1-4cf1-a1c7-3108e60b062e req-080b64e6-efa2-4eed-9a3c-759ba14be829 service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Received event network-vif-plugged-eb47f210-8e1e-4310-b314-33a489cf4325 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.285514] env[63355]: DEBUG oslo_concurrency.lockutils [req-283d51b8-80a1-4cf1-a1c7-3108e60b062e req-080b64e6-efa2-4eed-9a3c-759ba14be829 service nova] Acquiring lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.285870] env[63355]: DEBUG oslo_concurrency.lockutils [req-283d51b8-80a1-4cf1-a1c7-3108e60b062e req-080b64e6-efa2-4eed-9a3c-759ba14be829 service nova] Lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.003s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.286113] env[63355]: DEBUG oslo_concurrency.lockutils [req-283d51b8-80a1-4cf1-a1c7-3108e60b062e req-080b64e6-efa2-4eed-9a3c-759ba14be829 service nova] Lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.286322] env[63355]: DEBUG nova.compute.manager [req-283d51b8-80a1-4cf1-a1c7-3108e60b062e req-080b64e6-efa2-4eed-9a3c-759ba14be829 service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] No waiting events found dispatching network-vif-plugged-eb47f210-8e1e-4310-b314-33a489cf4325 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 974.286496] env[63355]: WARNING nova.compute.manager [req-283d51b8-80a1-4cf1-a1c7-3108e60b062e req-080b64e6-efa2-4eed-9a3c-759ba14be829 service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Received unexpected event network-vif-plugged-eb47f210-8e1e-4310-b314-33a489cf4325 for instance with vm_state building and task_state spawning. [ 974.287661] env[63355]: INFO nova.scheduler.client.report [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Deleted allocations for instance 0b660647-697a-445b-bca4-82ef6975ea75 [ 974.387726] env[63355]: DEBUG nova.network.neutron [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Successfully updated port: eb47f210-8e1e-4310-b314-33a489cf4325 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 974.550560] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349915, 'name': ReconfigVM_Task, 'duration_secs': 0.165451} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.551180] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Reconfigured VM instance instance-00000049 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 974.563858] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c0feda2-2ce2-4828-8f09-cfbe8be98c87 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.582586] env[63355]: DEBUG oslo_vmware.api [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349913, 'name': PowerOnVM_Task, 'duration_secs': 0.917771} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.582586] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.583643] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 974.583915] env[63355]: DEBUG nova.compute.manager [None req-ffd36c9d-131f-4816-8b2c-445a76d22532 tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.585091] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c037cf-5f60-49c4-9afe-eb255e9abb88 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.592025] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349916, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.593201] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for the task: (returnval){ [ 974.593201] env[63355]: value = "task-1349917" [ 974.593201] env[63355]: _type = "Task" [ 974.593201] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.614247] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349917, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.780120] env[63355]: DEBUG nova.network.neutron [-] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.782217] env[63355]: DEBUG nova.compute.manager [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 974.798263] env[63355]: DEBUG oslo_concurrency.lockutils [None req-749f6740-3b9f-4b4e-b74a-018efe097503 tempest-ListImageFiltersTestJSON-404199601 tempest-ListImageFiltersTestJSON-404199601-project-member] Lock "0b660647-697a-445b-bca4-82ef6975ea75" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.012s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.891094] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquiring lock "refresh_cache-c630c5fe-6907-4952-9807-6e59bd1cc9e1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.891094] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquired lock "refresh_cache-c630c5fe-6907-4952-9807-6e59bd1cc9e1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.891094] env[63355]: DEBUG nova.network.neutron [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 974.947767] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 974.949068] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b417554b-8ecd-4bfd-9e70-0285f9cb2323 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.959722] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 974.960624] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35966f5d-98af-40c1-8ef1-4213ed67cc14 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.078780] env[63355]: DEBUG oslo_vmware.api [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1349916, 'name': PowerOnVM_Task, 'duration_secs': 0.663575} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.080127] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.080623] env[63355]: INFO nova.compute.manager [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Took 9.03 seconds to spawn the instance on the hypervisor. [ 975.080836] env[63355]: DEBUG nova.compute.manager [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.082642] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653d956a-1af3-40e4-8a1c-144f734953bd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.106037] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 975.106037] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 975.106037] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleting the datastore file [datastore2] cd9de9b8-ad42-4dbb-b435-927738a55f3f {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.109827] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-126c52d4-c56b-412d-ba5e-44b64470f9b3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.126023] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349917, 'name': ReconfigVM_Task, 'duration_secs': 0.19428} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.126023] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287631', 'volume_id': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'name': 'volume-16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ecbacdb3-516e-43c3-96e3-4961b76565ca', 'attached_at': '', 'detached_at': '', 'volume_id': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1', 'serial': '16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 975.126023] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 975.126023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7911e22e-eab9-4648-ba95-69e70f83f0f2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.131053] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 975.131053] env[63355]: value = "task-1349919" [ 975.131053] env[63355]: _type = "Task" [ 975.131053] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.138290] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 975.139282] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3848c21-38b5-40c7-a1ff-287138e67939 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.144890] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349919, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.224039] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 975.224039] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 975.224039] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Deleting the datastore file [datastore1] ecbacdb3-516e-43c3-96e3-4961b76565ca {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.224039] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4110fa6b-5ed6-4544-b1b1-88a7f4b16e58 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.233575] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for the task: (returnval){ [ 975.233575] env[63355]: value = "task-1349921" [ 975.233575] env[63355]: _type = "Task" [ 975.233575] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.243132] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.287977] env[63355]: INFO nova.compute.manager [-] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Took 1.79 seconds to deallocate network for instance. [ 975.313195] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.435632] env[63355]: DEBUG nova.network.neutron [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 975.537019] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3560833e-ccab-4f00-905d-d75678b5f4d5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.545475] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d280f1-6436-4bfb-9759-619ec58ae9c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.575906] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4a80dc-399d-4dfd-a23e-c5f773d619ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.586646] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa86bde-4aa7-405f-9ed8-f7d5ffa29f94 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.612172] env[63355]: DEBUG nova.compute.provider_tree [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 975.618401] env[63355]: INFO nova.compute.manager [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Took 29.08 seconds to build instance. [ 975.626157] env[63355]: DEBUG nova.network.neutron [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Updating instance_info_cache with network_info: [{"id": "eb47f210-8e1e-4310-b314-33a489cf4325", "address": "fa:16:3e:3f:08:98", "network": {"id": "0590d30a-4e5f-425e-8793-16722bf317e6", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-2037043110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "226038844bda45b28033d6ea40a9bc9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb47f210-8e", "ovs_interfaceid": "eb47f210-8e1e-4310-b314-33a489cf4325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.644634] env[63355]: DEBUG oslo_vmware.api [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1349919, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.370391} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.645372] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 975.645595] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 975.645775] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 975.679104] env[63355]: INFO nova.scheduler.client.report [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted allocations for instance cd9de9b8-ad42-4dbb-b435-927738a55f3f [ 975.742192] env[63355]: DEBUG oslo_vmware.api [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Task: {'id': task-1349921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208589} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.742461] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 975.742707] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 975.742857] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 975.803930] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.819548] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 975.819905] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb2ef005-1d7f-4161-ad18-c79df0a11a7f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.829928] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5737a970-cb8f-4d7d-a35e-efd16b1601d0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.863046] env[63355]: ERROR nova.compute.manager [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Failed to detach volume 16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1 from /dev/sda: nova.exception.InstanceNotFound: Instance ecbacdb3-516e-43c3-96e3-4961b76565ca could not be found. [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Traceback (most recent call last): [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self.driver.rebuild(**kwargs) [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] raise NotImplementedError() [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] NotImplementedError [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] During handling of the above exception, another exception occurred: [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Traceback (most recent call last): [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self.driver.detach_volume(context, old_connection_info, [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] return self._volumeops.detach_volume(connection_info, instance) [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self._detach_volume_vmdk(connection_info, instance) [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] stable_ref.fetch_moref(session) [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] raise exception.InstanceNotFound(instance_id=self._uuid) [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] nova.exception.InstanceNotFound: Instance ecbacdb3-516e-43c3-96e3-4961b76565ca could not be found. [ 975.863046] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] [ 976.057642] env[63355]: DEBUG nova.compute.utils [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Build of instance ecbacdb3-516e-43c3-96e3-4961b76565ca aborted: Failed to rebuild volume backed instance. {{(pid=63355) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 976.061384] env[63355]: ERROR nova.compute.manager [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance ecbacdb3-516e-43c3-96e3-4961b76565ca aborted: Failed to rebuild volume backed instance. [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Traceback (most recent call last): [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self.driver.rebuild(**kwargs) [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] raise NotImplementedError() [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] NotImplementedError [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] During handling of the above exception, another exception occurred: [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Traceback (most recent call last): [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self._detach_root_volume(context, instance, root_bdm) [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] with excutils.save_and_reraise_exception(): [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self.force_reraise() [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] raise self.value [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self.driver.detach_volume(context, old_connection_info, [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] return self._volumeops.detach_volume(connection_info, instance) [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self._detach_volume_vmdk(connection_info, instance) [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] stable_ref.fetch_moref(session) [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] raise exception.InstanceNotFound(instance_id=self._uuid) [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] nova.exception.InstanceNotFound: Instance ecbacdb3-516e-43c3-96e3-4961b76565ca could not be found. [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] During handling of the above exception, another exception occurred: [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Traceback (most recent call last): [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] yield [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 976.061384] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self._do_rebuild_instance_with_claim( [ 976.062606] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 976.062606] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self._do_rebuild_instance( [ 976.062606] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 976.062606] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self._rebuild_default_impl(**kwargs) [ 976.062606] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 976.062606] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] self._rebuild_volume_backed_instance( [ 976.062606] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 976.062606] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] raise exception.BuildAbortException( [ 976.062606] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] nova.exception.BuildAbortException: Build of instance ecbacdb3-516e-43c3-96e3-4961b76565ca aborted: Failed to rebuild volume backed instance. [ 976.062606] env[63355]: ERROR nova.compute.manager [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] [ 976.123526] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "8e9ff9b7-636f-48df-9168-509d733278f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.123526] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e9ff9b7-636f-48df-9168-509d733278f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.123755] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "8e9ff9b7-636f-48df-9168-509d733278f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.123853] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e9ff9b7-636f-48df-9168-509d733278f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.124038] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e9ff9b7-636f-48df-9168-509d733278f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.127044] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c5363f5e-6d7b-4436-a15f-5e271affa893 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.596s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.127913] env[63355]: INFO nova.compute.manager [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Terminating instance [ 976.129602] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Releasing lock "refresh_cache-c630c5fe-6907-4952-9807-6e59bd1cc9e1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.130217] env[63355]: DEBUG nova.compute.manager [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Instance network_info: |[{"id": "eb47f210-8e1e-4310-b314-33a489cf4325", "address": "fa:16:3e:3f:08:98", "network": {"id": "0590d30a-4e5f-425e-8793-16722bf317e6", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-2037043110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "226038844bda45b28033d6ea40a9bc9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb47f210-8e", "ovs_interfaceid": "eb47f210-8e1e-4310-b314-33a489cf4325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 976.130586] env[63355]: DEBUG nova.compute.manager [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.131275] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 976.132701] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:08:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb47f210-8e1e-4310-b314-33a489cf4325', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 976.140459] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Creating folder: Project (226038844bda45b28033d6ea40a9bc9b). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 976.141336] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe95a803-457f-4447-8efc-7f8bb0400170 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.148032] env[63355]: ERROR nova.scheduler.client.report [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [req-d5b094c1-69e2-4b9f-bb3f-4b64180e878c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 47abb610-db7e-4770-911d-187dd075ef8b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d5b094c1-69e2-4b9f-bb3f-4b64180e878c"}]} [ 976.148865] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1594f917-6f9d-4e32-9f30-fe6f766f1b78 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.158735] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 976.159061] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9875d1ad-df28-4de4-a5f3-b828aa2650f6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.162721] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Created folder: Project (226038844bda45b28033d6ea40a9bc9b) in parent group-v287607. [ 976.162938] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Creating folder: Instances. Parent ref: group-v287719. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 976.164135] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-947b9b15-f658-46b5-92c5-3d0b892ff74f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.167911] env[63355]: DEBUG oslo_vmware.api [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 976.167911] env[63355]: value = "task-1349923" [ 976.167911] env[63355]: _type = "Task" [ 976.167911] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.172652] env[63355]: DEBUG nova.scheduler.client.report [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Refreshing inventories for resource provider 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 976.176521] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Created folder: Instances in parent group-v287719. [ 976.176821] env[63355]: DEBUG oslo.service.loopingcall [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.177153] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 976.177486] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d1634e5-ba62-4768-b468-eac922e00e85 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.201296] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.201895] env[63355]: DEBUG oslo_vmware.api [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349923, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.210021] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 976.210021] env[63355]: value = "task-1349925" [ 976.210021] env[63355]: _type = "Task" [ 976.210021] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.217603] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349925, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.287433] env[63355]: DEBUG nova.scheduler.client.report [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Updating ProviderTree inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 976.287433] env[63355]: DEBUG nova.compute.provider_tree [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 976.300511] env[63355]: DEBUG nova.scheduler.client.report [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Refreshing aggregate associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, aggregates: None {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 976.320054] env[63355]: DEBUG nova.scheduler.client.report [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Refreshing trait associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 976.458945] env[63355]: DEBUG nova.compute.manager [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Received event network-vif-deleted-72ce124a-30e1-4794-8c3f-3156b6ce619c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.459436] env[63355]: DEBUG nova.compute.manager [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Received event network-changed-eb47f210-8e1e-4310-b314-33a489cf4325 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.459779] env[63355]: DEBUG nova.compute.manager [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Refreshing instance network info cache due to event network-changed-eb47f210-8e1e-4310-b314-33a489cf4325. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 976.460201] env[63355]: DEBUG oslo_concurrency.lockutils [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] Acquiring lock "refresh_cache-c630c5fe-6907-4952-9807-6e59bd1cc9e1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.460527] env[63355]: DEBUG oslo_concurrency.lockutils [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] Acquired lock "refresh_cache-c630c5fe-6907-4952-9807-6e59bd1cc9e1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.462292] env[63355]: DEBUG nova.network.neutron [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Refreshing network info cache for port eb47f210-8e1e-4310-b314-33a489cf4325 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 976.548461] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6cef672-38f0-4e53-afe6-cadc9a563edf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.556973] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3cf16d-6d66-4c29-b545-831853a1fb08 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.597062] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbfec72a-271d-4e41-ae90-d2367cae6b7e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.610197] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630c2a15-152d-46a5-8b1f-7bf37084257f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.629000] env[63355]: DEBUG nova.compute.provider_tree [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 976.653334] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.680635] env[63355]: DEBUG oslo_vmware.api [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349923, 'name': PowerOffVM_Task, 'duration_secs': 0.262629} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.680937] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 976.681129] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 976.681399] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a5eb773-b27d-462e-b5dc-e88b7be2732e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.718076] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349925, 'name': CreateVM_Task, 'duration_secs': 0.429538} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.718315] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 976.719070] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.719261] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.719595] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 976.719864] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68bf416f-77a4-4d64-835f-a6c0ed88e5fc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.725067] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for the task: (returnval){ [ 976.725067] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528e08f5-8fce-c3d4-18c2-903191cd29ab" [ 976.725067] env[63355]: _type = "Task" [ 976.725067] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.737023] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528e08f5-8fce-c3d4-18c2-903191cd29ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.743259] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 976.743477] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 976.743655] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Deleting the datastore file [datastore1] 8e9ff9b7-636f-48df-9168-509d733278f8 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 976.743925] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78cf211d-0f78-42cf-8c04-c9ab7a34939d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.751045] env[63355]: DEBUG oslo_vmware.api [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 976.751045] env[63355]: value = "task-1349927" [ 976.751045] env[63355]: _type = "Task" [ 976.751045] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.761594] env[63355]: DEBUG oslo_vmware.api [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.186195] env[63355]: DEBUG nova.scheduler.client.report [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Updated inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b with generation 103 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 977.186503] env[63355]: DEBUG nova.compute.provider_tree [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Updating resource provider 47abb610-db7e-4770-911d-187dd075ef8b generation from 103 to 104 during operation: update_inventory {{(pid=63355) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 977.186685] env[63355]: DEBUG nova.compute.provider_tree [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 977.235119] env[63355]: DEBUG nova.compute.manager [req-8660bc3d-dd89-4842-90c5-12af6dc65c25 req-eb0447fe-1abc-4d37-b953-3c7109a0fe02 service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Received event network-changed-d4c4170a-d27a-4f64-8789-6e01e1adb3df {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.235119] env[63355]: DEBUG nova.compute.manager [req-8660bc3d-dd89-4842-90c5-12af6dc65c25 req-eb0447fe-1abc-4d37-b953-3c7109a0fe02 service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Refreshing instance network info cache due to event network-changed-d4c4170a-d27a-4f64-8789-6e01e1adb3df. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 977.235119] env[63355]: DEBUG oslo_concurrency.lockutils [req-8660bc3d-dd89-4842-90c5-12af6dc65c25 req-eb0447fe-1abc-4d37-b953-3c7109a0fe02 service nova] Acquiring lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.235119] env[63355]: DEBUG oslo_concurrency.lockutils [req-8660bc3d-dd89-4842-90c5-12af6dc65c25 req-eb0447fe-1abc-4d37-b953-3c7109a0fe02 service nova] Acquired lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.235119] env[63355]: DEBUG nova.network.neutron [req-8660bc3d-dd89-4842-90c5-12af6dc65c25 req-eb0447fe-1abc-4d37-b953-3c7109a0fe02 service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Refreshing network info cache for port d4c4170a-d27a-4f64-8789-6e01e1adb3df {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.242318] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528e08f5-8fce-c3d4-18c2-903191cd29ab, 'name': SearchDatastore_Task, 'duration_secs': 0.015748} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.245722] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.246102] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.246428] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.246655] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.246891] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.247692] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6242fd71-a1fe-4d39-a8a3-12253f7b14e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.258407] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.258654] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 977.262692] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a5f143e-f918-49c0-a57d-eb595dc60e5d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.265508] env[63355]: DEBUG oslo_vmware.api [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349927, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.250014} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.266276] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 977.266559] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 977.266804] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 977.267081] env[63355]: INFO nova.compute.manager [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 977.267402] env[63355]: DEBUG oslo.service.loopingcall [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.268102] env[63355]: DEBUG nova.compute.manager [-] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 977.268746] env[63355]: DEBUG nova.network.neutron [-] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 977.271703] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for the task: (returnval){ [ 977.271703] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521b7e2c-f57f-e630-07fc-d893730826b0" [ 977.271703] env[63355]: _type = "Task" [ 977.271703] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.282058] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521b7e2c-f57f-e630-07fc-d893730826b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.308165] env[63355]: DEBUG nova.network.neutron [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Updated VIF entry in instance network info cache for port eb47f210-8e1e-4310-b314-33a489cf4325. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 977.308165] env[63355]: DEBUG nova.network.neutron [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Updating instance_info_cache with network_info: [{"id": "eb47f210-8e1e-4310-b314-33a489cf4325", "address": "fa:16:3e:3f:08:98", "network": {"id": "0590d30a-4e5f-425e-8793-16722bf317e6", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-2037043110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "226038844bda45b28033d6ea40a9bc9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb47f210-8e", "ovs_interfaceid": "eb47f210-8e1e-4310-b314-33a489cf4325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.694801] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.454s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.695308] env[63355]: DEBUG nova.compute.manager [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 977.697712] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.799s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.699273] env[63355]: INFO nova.compute.claims [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.783725] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521b7e2c-f57f-e630-07fc-d893730826b0, 'name': SearchDatastore_Task, 'duration_secs': 0.016471} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.784708] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aead0b92-99ea-4372-aa16-8f73a0d19798 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.792019] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for the task: (returnval){ [ 977.792019] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525d4dc5-8ad2-dde3-94a7-650edd47bce2" [ 977.792019] env[63355]: _type = "Task" [ 977.792019] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.801069] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525d4dc5-8ad2-dde3-94a7-650edd47bce2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.812763] env[63355]: DEBUG oslo_concurrency.lockutils [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] Releasing lock "refresh_cache-c630c5fe-6907-4952-9807-6e59bd1cc9e1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.813056] env[63355]: DEBUG nova.compute.manager [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Received event network-vif-unplugged-850ef0bb-ad5f-4627-b482-a78817b84343 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.813430] env[63355]: DEBUG oslo_concurrency.lockutils [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] Acquiring lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.813497] env[63355]: DEBUG oslo_concurrency.lockutils [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.813656] env[63355]: DEBUG oslo_concurrency.lockutils [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.814289] env[63355]: DEBUG nova.compute.manager [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] No waiting events found dispatching network-vif-unplugged-850ef0bb-ad5f-4627-b482-a78817b84343 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 977.814289] env[63355]: DEBUG nova.compute.manager [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Received event network-vif-unplugged-850ef0bb-ad5f-4627-b482-a78817b84343 for instance with task_state deleting. {{(pid=63355) _process_instance_event /opt/stack/nova/nova/compute/manager.py:10909}} [ 977.814289] env[63355]: DEBUG nova.compute.manager [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Received event network-changed-850ef0bb-ad5f-4627-b482-a78817b84343 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.814457] env[63355]: DEBUG nova.compute.manager [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Refreshing instance network info cache due to event network-changed-850ef0bb-ad5f-4627-b482-a78817b84343. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 977.814536] env[63355]: DEBUG oslo_concurrency.lockutils [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] Acquiring lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.814669] env[63355]: DEBUG oslo_concurrency.lockutils [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] Acquired lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.814826] env[63355]: DEBUG nova.network.neutron [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Refreshing network info cache for port 850ef0bb-ad5f-4627-b482-a78817b84343 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.976421] env[63355]: DEBUG nova.network.neutron [req-8660bc3d-dd89-4842-90c5-12af6dc65c25 req-eb0447fe-1abc-4d37-b953-3c7109a0fe02 service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updated VIF entry in instance network info cache for port d4c4170a-d27a-4f64-8789-6e01e1adb3df. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 977.976846] env[63355]: DEBUG nova.network.neutron [req-8660bc3d-dd89-4842-90c5-12af6dc65c25 req-eb0447fe-1abc-4d37-b953-3c7109a0fe02 service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updating instance_info_cache with network_info: [{"id": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "address": "fa:16:3e:fb:af:dc", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4c4170a-d2", "ovs_interfaceid": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.062783] env[63355]: DEBUG nova.network.neutron [-] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.084948] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquiring lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.085212] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.109146] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.203713] env[63355]: DEBUG nova.compute.utils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.207355] env[63355]: DEBUG nova.compute.manager [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 978.207355] env[63355]: DEBUG nova.network.neutron [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 978.257981] env[63355]: DEBUG nova.policy [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bf70ebeeb3a4327ada51d18f5651476', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce0682a99ac94aeea463c961b84e6b58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.301575] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525d4dc5-8ad2-dde3-94a7-650edd47bce2, 'name': SearchDatastore_Task, 'duration_secs': 0.018128} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.301911] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.302255] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] c630c5fe-6907-4952-9807-6e59bd1cc9e1/c630c5fe-6907-4952-9807-6e59bd1cc9e1.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 978.302565] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30347378-b5f6-4339-8db4-e104089f0819 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.309361] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for the task: (returnval){ [ 978.309361] env[63355]: value = "task-1349928" [ 978.309361] env[63355]: _type = "Task" [ 978.309361] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.319090] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349928, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.384141] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Acquiring lock "ecbacdb3-516e-43c3-96e3-4961b76565ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.384592] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "ecbacdb3-516e-43c3-96e3-4961b76565ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.384810] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Acquiring lock "ecbacdb3-516e-43c3-96e3-4961b76565ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.385057] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "ecbacdb3-516e-43c3-96e3-4961b76565ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.385355] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "ecbacdb3-516e-43c3-96e3-4961b76565ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.388080] env[63355]: INFO nova.compute.manager [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Terminating instance [ 978.390803] env[63355]: DEBUG nova.compute.manager [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 978.391155] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ce0cb01-b521-490f-96f9-82244ed56e12 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.401874] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cef564e-4f99-4577-8947-bb248721119b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.430797] env[63355]: WARNING nova.virt.vmwareapi.driver [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance ecbacdb3-516e-43c3-96e3-4961b76565ca could not be found. [ 978.431062] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.431853] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-004bc0c2-6eba-4bd4-b911-11bef1f04a2c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.439926] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d13fbe4-6452-4fc0-907b-41bf09dd935f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.469179] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ecbacdb3-516e-43c3-96e3-4961b76565ca could not be found. [ 978.469419] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 978.469609] env[63355]: INFO nova.compute.manager [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Took 0.08 seconds to destroy the instance on the hypervisor. [ 978.469913] env[63355]: DEBUG oslo.service.loopingcall [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.470166] env[63355]: DEBUG nova.compute.manager [-] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 978.470263] env[63355]: DEBUG nova.network.neutron [-] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 978.480228] env[63355]: DEBUG oslo_concurrency.lockutils [req-8660bc3d-dd89-4842-90c5-12af6dc65c25 req-eb0447fe-1abc-4d37-b953-3c7109a0fe02 service nova] Releasing lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.489458] env[63355]: DEBUG nova.compute.manager [req-c2f72777-4918-41b7-962b-cb4c32df4dda req-2927c577-bd05-406f-9b95-daa2c6379447 service nova] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Received event network-vif-deleted-a0ceb2ca-3586-4fbf-9f0f-3a568bba6f3a {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.565145] env[63355]: INFO nova.compute.manager [-] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Took 1.30 seconds to deallocate network for instance. [ 978.587869] env[63355]: DEBUG nova.compute.manager [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 978.663451] env[63355]: DEBUG nova.network.neutron [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Updated VIF entry in instance network info cache for port 850ef0bb-ad5f-4627-b482-a78817b84343. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.663845] env[63355]: DEBUG nova.network.neutron [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Updating instance_info_cache with network_info: [{"id": "850ef0bb-ad5f-4627-b482-a78817b84343", "address": "fa:16:3e:42:ee:f7", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": null, "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap850ef0bb-ad", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.683514] env[63355]: DEBUG nova.network.neutron [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Successfully created port: 52d11eb4-329c-4989-9498-377fd46350cb {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.708703] env[63355]: DEBUG nova.compute.manager [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 978.827839] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349928, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.991039] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475f26d2-c344-493a-9757-8494b63619af {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.999692] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7dfc4b-aec8-45d7-88e8-70b1bf206561 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.037807] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd267a7-9a5d-49a1-b668-a84042ea5e75 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.045902] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512d0d24-baa0-4edd-bef5-63761ee24230 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.062695] env[63355]: DEBUG nova.compute.provider_tree [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.072625] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.117774] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.167384] env[63355]: DEBUG oslo_concurrency.lockutils [req-ef6a5b8f-2eed-45e4-bdc6-aea0508c3c40 req-77c9bb30-d03f-4606-9c57-ca20c41065ab service nova] Releasing lock "refresh_cache-cd9de9b8-ad42-4dbb-b435-927738a55f3f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.322146] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349928, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.717813} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.322457] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] c630c5fe-6907-4952-9807-6e59bd1cc9e1/c630c5fe-6907-4952-9807-6e59bd1cc9e1.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 979.322850] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.323253] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7c07419-2045-4596-af5a-88f3d2263617 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.330552] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for the task: (returnval){ [ 979.330552] env[63355]: value = "task-1349929" [ 979.330552] env[63355]: _type = "Task" [ 979.330552] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.339257] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349929, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.510275] env[63355]: DEBUG nova.compute.manager [req-32a38c9f-27b4-4259-8418-104f95c45bb6 req-843d3e5d-ca8d-410d-a70b-63bdc8667453 service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Received event network-vif-deleted-0500cf25-c654-4748-a58a-8e51bd919768 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.510420] env[63355]: INFO nova.compute.manager [req-32a38c9f-27b4-4259-8418-104f95c45bb6 req-843d3e5d-ca8d-410d-a70b-63bdc8667453 service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Neutron deleted interface 0500cf25-c654-4748-a58a-8e51bd919768; detaching it from the instance and deleting it from the info cache [ 979.510626] env[63355]: DEBUG nova.network.neutron [req-32a38c9f-27b4-4259-8418-104f95c45bb6 req-843d3e5d-ca8d-410d-a70b-63bdc8667453 service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.565435] env[63355]: DEBUG nova.scheduler.client.report [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.728060] env[63355]: DEBUG nova.compute.manager [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 979.755732] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.755995] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.756276] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.756498] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.756668] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.756833] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.758137] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.758375] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.758572] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.758754] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.758941] env[63355]: DEBUG nova.virt.hardware [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.761533] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7232be17-9ad6-4092-9f1e-9751e6dd2ef3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.768107] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3844bac-8a06-4f0b-a87f-d11114dbcf95 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.840719] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349929, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070638} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.840951] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 979.841749] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5f6cdb-c997-4568-994e-9c1716b16376 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.866115] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] c630c5fe-6907-4952-9807-6e59bd1cc9e1/c630c5fe-6907-4952-9807-6e59bd1cc9e1.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 979.866476] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-489b5b3c-2bcd-4ab6-9d06-58965313a001 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.886815] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for the task: (returnval){ [ 979.886815] env[63355]: value = "task-1349930" [ 979.886815] env[63355]: _type = "Task" [ 979.886815] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.894821] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349930, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.984501] env[63355]: DEBUG nova.network.neutron [-] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.017307] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-437fe8a3-2e73-4bab-8a1b-469c3e7a9c83 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.030028] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6341027d-14ed-4b4c-87c2-0862f818a0e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.060378] env[63355]: DEBUG nova.compute.manager [req-32a38c9f-27b4-4259-8418-104f95c45bb6 req-843d3e5d-ca8d-410d-a70b-63bdc8667453 service nova] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Detach interface failed, port_id=0500cf25-c654-4748-a58a-8e51bd919768, reason: Instance ecbacdb3-516e-43c3-96e3-4961b76565ca could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 980.070696] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.071326] env[63355]: DEBUG nova.compute.manager [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 980.075926] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.124s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.076258] env[63355]: DEBUG nova.objects.instance [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lazy-loading 'resources' on Instance uuid 1570cb36-76e0-4d06-8080-735b5246e92e {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.397503] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349930, 'name': ReconfigVM_Task, 'duration_secs': 0.463282} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.397800] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Reconfigured VM instance instance-00000054 to attach disk [datastore1] c630c5fe-6907-4952-9807-6e59bd1cc9e1/c630c5fe-6907-4952-9807-6e59bd1cc9e1.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.398507] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddc0426b-dbd0-4062-be2d-4c283bb51062 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.405650] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for the task: (returnval){ [ 980.405650] env[63355]: value = "task-1349931" [ 980.405650] env[63355]: _type = "Task" [ 980.405650] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.414368] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349931, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.424830] env[63355]: DEBUG oslo_vmware.rw_handles [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52205b72-7bf7-a082-bdae-de171644ed4c/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 980.425639] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a980a6-936a-412d-a039-0f10c47d58de {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.431512] env[63355]: DEBUG oslo_vmware.rw_handles [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52205b72-7bf7-a082-bdae-de171644ed4c/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 980.431685] env[63355]: ERROR oslo_vmware.rw_handles [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52205b72-7bf7-a082-bdae-de171644ed4c/disk-0.vmdk due to incomplete transfer. [ 980.431914] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c1baed19-ad1b-43dc-aa4c-e18bf1e5f5ba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.437496] env[63355]: DEBUG oslo_vmware.rw_handles [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52205b72-7bf7-a082-bdae-de171644ed4c/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 980.437700] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Uploaded image eabfdc0f-53a1-4aa1-a863-de4d0008a5e5 to the Glance image server {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 980.439925] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 980.440185] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-70a8dbfc-c1f4-4a9e-936f-5daf62a642f4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.444772] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 980.444772] env[63355]: value = "task-1349932" [ 980.444772] env[63355]: _type = "Task" [ 980.444772] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.452216] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349932, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.477678] env[63355]: DEBUG nova.network.neutron [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Successfully updated port: 52d11eb4-329c-4989-9498-377fd46350cb {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.487824] env[63355]: INFO nova.compute.manager [-] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Took 2.02 seconds to deallocate network for instance. [ 980.523715] env[63355]: DEBUG nova.compute.manager [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Received event network-vif-plugged-52d11eb4-329c-4989-9498-377fd46350cb {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.523953] env[63355]: DEBUG oslo_concurrency.lockutils [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] Acquiring lock "fd55ddfc-f376-4e15-961e-6826f31a7890-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.524229] env[63355]: DEBUG oslo_concurrency.lockutils [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.524421] env[63355]: DEBUG oslo_concurrency.lockutils [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.524581] env[63355]: DEBUG nova.compute.manager [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] No waiting events found dispatching network-vif-plugged-52d11eb4-329c-4989-9498-377fd46350cb {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 980.524794] env[63355]: WARNING nova.compute.manager [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Received unexpected event network-vif-plugged-52d11eb4-329c-4989-9498-377fd46350cb for instance with vm_state building and task_state spawning. [ 980.524895] env[63355]: DEBUG nova.compute.manager [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Received event network-changed-52d11eb4-329c-4989-9498-377fd46350cb {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.526213] env[63355]: DEBUG nova.compute.manager [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Refreshing instance network info cache due to event network-changed-52d11eb4-329c-4989-9498-377fd46350cb. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 980.526477] env[63355]: DEBUG oslo_concurrency.lockutils [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] Acquiring lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.526636] env[63355]: DEBUG oslo_concurrency.lockutils [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] Acquired lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.526807] env[63355]: DEBUG nova.network.neutron [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Refreshing network info cache for port 52d11eb4-329c-4989-9498-377fd46350cb {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 980.580524] env[63355]: DEBUG nova.compute.utils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 980.586169] env[63355]: DEBUG nova.compute.manager [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 980.586169] env[63355]: DEBUG nova.network.neutron [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 980.647400] env[63355]: DEBUG nova.policy [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3721c4a73df54ae9b44110cb9e8590a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc547065748241e8ac7b6c499ddaea66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 980.799429] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f41620-6f5a-4620-8b32-3b65da0200af {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.807285] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3579c2a2-5383-4453-ac17-4b1f9c90e0ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.838576] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35180897-40f6-4308-9ba8-54bdea46fc0d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.846210] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3aa67c0-986c-4df2-800f-614cbba80d71 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.859314] env[63355]: DEBUG nova.compute.provider_tree [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.915438] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349931, 'name': Rename_Task, 'duration_secs': 0.168238} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.915718] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 980.916028] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-147a8608-314f-4ca5-8e15-5b1604bfbb8b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.921509] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for the task: (returnval){ [ 980.921509] env[63355]: value = "task-1349933" [ 980.921509] env[63355]: _type = "Task" [ 980.921509] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.928630] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349933, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.953787] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349932, 'name': Destroy_Task, 'duration_secs': 0.329727} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.954117] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Destroyed the VM [ 980.954382] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 980.954650] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-03622709-8c44-4dab-b137-e9ba3f07012e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.963277] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 980.963277] env[63355]: value = "task-1349934" [ 980.963277] env[63355]: _type = "Task" [ 980.963277] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.972962] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349934, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.981068] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.984388] env[63355]: DEBUG nova.network.neutron [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Successfully created port: 677de229-5411-464b-9680-0080e5794772 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.043512] env[63355]: INFO nova.compute.manager [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Took 0.56 seconds to detach 1 volumes for instance. [ 981.048800] env[63355]: DEBUG nova.compute.manager [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Deleting volume: 16f3c32f-0f59-4dbe-94e3-ff81c08f6cb1 {{(pid=63355) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 981.075083] env[63355]: DEBUG nova.network.neutron [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 981.088294] env[63355]: DEBUG nova.compute.manager [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.210680] env[63355]: DEBUG nova.network.neutron [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.362671] env[63355]: DEBUG nova.scheduler.client.report [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.441642] env[63355]: DEBUG oslo_vmware.api [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349933, 'name': PowerOnVM_Task, 'duration_secs': 0.509685} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.442276] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 981.442622] env[63355]: INFO nova.compute.manager [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Took 8.13 seconds to spawn the instance on the hypervisor. [ 981.444179] env[63355]: DEBUG nova.compute.manager [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 981.445075] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb79ad6-4776-4a12-b113-5c068abaed39 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.472941] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349934, 'name': RemoveSnapshot_Task} progress is 26%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.596801] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.713245] env[63355]: DEBUG oslo_concurrency.lockutils [req-0cb986ab-16d4-489c-b108-03db23f1a6c2 req-ae937c5a-ecb5-42a8-8678-9d2f1886c2ca service nova] Releasing lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.713672] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.713840] env[63355]: DEBUG nova.network.neutron [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.867838] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.875024] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.158s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.875024] env[63355]: INFO nova.compute.claims [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.895382] env[63355]: INFO nova.scheduler.client.report [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted allocations for instance 1570cb36-76e0-4d06-8080-735b5246e92e [ 981.969372] env[63355]: INFO nova.compute.manager [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Took 17.39 seconds to build instance. [ 981.977198] env[63355]: DEBUG oslo_vmware.api [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349934, 'name': RemoveSnapshot_Task, 'duration_secs': 0.976451} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.977485] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 981.977747] env[63355]: INFO nova.compute.manager [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Took 15.46 seconds to snapshot the instance on the hypervisor. [ 982.103119] env[63355]: DEBUG nova.compute.manager [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 982.128048] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 982.128333] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 982.128497] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 982.128683] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 982.128833] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 982.128984] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 982.129214] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 982.129382] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 982.129556] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 982.129723] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 982.129899] env[63355]: DEBUG nova.virt.hardware [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 982.130763] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c7022e-4c9e-45ba-979e-d62cc29c18e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.138943] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481a909e-d6f6-4e2b-9c26-9601209a2543 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.255222] env[63355]: DEBUG nova.network.neutron [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 982.402632] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6f635bde-169d-4b75-93a1-51f9a697d9b1 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "1570cb36-76e0-4d06-8080-735b5246e92e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.798s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.404975] env[63355]: DEBUG nova.network.neutron [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance_info_cache with network_info: [{"id": "52d11eb4-329c-4989-9498-377fd46350cb", "address": "fa:16:3e:64:c5:93", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52d11eb4-32", "ovs_interfaceid": "52d11eb4-329c-4989-9498-377fd46350cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.472962] env[63355]: DEBUG oslo_concurrency.lockutils [None req-786c4436-ded6-43d5-9902-a55eeb3aad37 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.913s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.520348] env[63355]: DEBUG nova.network.neutron [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Successfully updated port: 677de229-5411-464b-9680-0080e5794772 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 982.533823] env[63355]: DEBUG nova.compute.manager [None req-170c03db-4f03-4a55-bf3d-df0c40855f76 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Found 1 images (rotation: 2) {{(pid=63355) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 982.581254] env[63355]: DEBUG nova.compute.manager [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Received event network-vif-plugged-677de229-5411-464b-9680-0080e5794772 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.581469] env[63355]: DEBUG oslo_concurrency.lockutils [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] Acquiring lock "edb14fe7-d444-4fef-8c5d-d5616676e326-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.581678] env[63355]: DEBUG oslo_concurrency.lockutils [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] Lock "edb14fe7-d444-4fef-8c5d-d5616676e326-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.581851] env[63355]: DEBUG oslo_concurrency.lockutils [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] Lock "edb14fe7-d444-4fef-8c5d-d5616676e326-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.582032] env[63355]: DEBUG nova.compute.manager [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] No waiting events found dispatching network-vif-plugged-677de229-5411-464b-9680-0080e5794772 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 982.584059] env[63355]: WARNING nova.compute.manager [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Received unexpected event network-vif-plugged-677de229-5411-464b-9680-0080e5794772 for instance with vm_state building and task_state spawning. [ 982.584059] env[63355]: DEBUG nova.compute.manager [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Received event network-changed-677de229-5411-464b-9680-0080e5794772 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.584059] env[63355]: DEBUG nova.compute.manager [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Refreshing instance network info cache due to event network-changed-677de229-5411-464b-9680-0080e5794772. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.584059] env[63355]: DEBUG oslo_concurrency.lockutils [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] Acquiring lock "refresh_cache-edb14fe7-d444-4fef-8c5d-d5616676e326" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.584059] env[63355]: DEBUG oslo_concurrency.lockutils [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] Acquired lock "refresh_cache-edb14fe7-d444-4fef-8c5d-d5616676e326" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.584059] env[63355]: DEBUG nova.network.neutron [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Refreshing network info cache for port 677de229-5411-464b-9680-0080e5794772 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.906642] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.906929] env[63355]: DEBUG nova.compute.manager [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Instance network_info: |[{"id": "52d11eb4-329c-4989-9498-377fd46350cb", "address": "fa:16:3e:64:c5:93", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52d11eb4-32", "ovs_interfaceid": "52d11eb4-329c-4989-9498-377fd46350cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.907528] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:c5:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '130387c4-e4ec-4d95-8e9d-bb079baabad8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52d11eb4-329c-4989-9498-377fd46350cb', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.915472] env[63355]: DEBUG oslo.service.loopingcall [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.918297] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.919499] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea416773-1d5a-4521-94c1-37b0c36b3814 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.942733] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.942733] env[63355]: value = "task-1349936" [ 982.942733] env[63355]: _type = "Task" [ 982.942733] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.955413] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349936, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.023182] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "refresh_cache-edb14fe7-d444-4fef-8c5d-d5616676e326" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.068765] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae319c2-28e5-47fe-bc51-bec95a49af42 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.077595] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fd9545-24e0-4b46-a99c-1145b6dbc291 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.111300] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5114f3e-3539-47e8-8c91-5f333a48acdb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.119256] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0799d483-c093-4d12-bc82-ac217071d75f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.133228] env[63355]: DEBUG nova.compute.provider_tree [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.148672] env[63355]: DEBUG nova.network.neutron [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 983.151782] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquiring lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.151782] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.152065] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquiring lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.152280] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.152458] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.155756] env[63355]: INFO nova.compute.manager [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Terminating instance [ 983.157713] env[63355]: DEBUG nova.compute.manager [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 983.157994] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 983.158868] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e4746f6-3d88-459b-8146-e69367a60971 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.169195] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 983.169582] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-305255f3-9d01-44af-804c-2e4596b2c47b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.177652] env[63355]: DEBUG oslo_vmware.api [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for the task: (returnval){ [ 983.177652] env[63355]: value = "task-1349937" [ 983.177652] env[63355]: _type = "Task" [ 983.177652] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.194356] env[63355]: DEBUG oslo_vmware.api [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.242890] env[63355]: DEBUG nova.network.neutron [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.408631] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.409016] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.409250] env[63355]: INFO nova.compute.manager [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Shelving [ 983.453572] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349936, 'name': CreateVM_Task, 'duration_secs': 0.349197} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.453759] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 983.454744] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.454987] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.455259] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.455567] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0756390-9594-4cb5-9155-ccd5ea8c92db {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.460068] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 983.460068] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522546eb-af94-4fdd-2090-affdc1991128" [ 983.460068] env[63355]: _type = "Task" [ 983.460068] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.468135] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522546eb-af94-4fdd-2090-affdc1991128, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.637142] env[63355]: DEBUG nova.scheduler.client.report [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.688134] env[63355]: DEBUG oslo_vmware.api [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349937, 'name': PowerOffVM_Task, 'duration_secs': 0.210842} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.688425] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.688593] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 983.688851] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d3f8ac4-4f23-430e-83ae-0ad821421ed7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.746807] env[63355]: DEBUG oslo_concurrency.lockutils [req-47ec9d95-315b-4a22-9a1f-522aff266af7 req-66943e3a-edfa-4d82-875e-47f09e672df9 service nova] Releasing lock "refresh_cache-edb14fe7-d444-4fef-8c5d-d5616676e326" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.747235] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "refresh_cache-edb14fe7-d444-4fef-8c5d-d5616676e326" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.747407] env[63355]: DEBUG nova.network.neutron [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 983.758274] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 983.758495] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 983.758680] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Deleting the datastore file [datastore1] c630c5fe-6907-4952-9807-6e59bd1cc9e1 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.758940] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d05cd35-5912-45c6-8e4c-f05df3c9b5a2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.766015] env[63355]: DEBUG oslo_vmware.api [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for the task: (returnval){ [ 983.766015] env[63355]: value = "task-1349939" [ 983.766015] env[63355]: _type = "Task" [ 983.766015] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.773643] env[63355]: DEBUG oslo_vmware.api [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349939, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.824296] env[63355]: DEBUG nova.compute.manager [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.825241] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c03a7f-47b0-44f5-ae92-aa450ded91b3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.919413] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 983.919413] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-122ea269-6f43-4296-9f50-1e5386802501 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.927131] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 983.927131] env[63355]: value = "task-1349940" [ 983.927131] env[63355]: _type = "Task" [ 983.927131] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.935427] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349940, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.970670] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522546eb-af94-4fdd-2090-affdc1991128, 'name': SearchDatastore_Task, 'duration_secs': 0.009683} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.970997] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.971258] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.971507] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.971668] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.971839] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.972317] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1f3b400-d9f2-425b-84bf-4dc30031a60a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.980014] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.980216] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.980926] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7515b86e-2a5f-44f2-bc27-ee25365fb2d7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.986260] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 983.986260] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521673c4-b48b-cad0-6cc8-8f1941865d18" [ 983.986260] env[63355]: _type = "Task" [ 983.986260] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.994549] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521673c4-b48b-cad0-6cc8-8f1941865d18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.144440] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.144971] env[63355]: DEBUG nova.compute.manager [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 984.147809] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.835s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.149124] env[63355]: INFO nova.compute.claims [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.275777] env[63355]: DEBUG oslo_vmware.api [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Task: {'id': task-1349939, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124884} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.276535] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 984.276535] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 984.276664] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 984.276852] env[63355]: INFO nova.compute.manager [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 984.277081] env[63355]: DEBUG oslo.service.loopingcall [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.277299] env[63355]: DEBUG nova.compute.manager [-] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 984.277390] env[63355]: DEBUG nova.network.neutron [-] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 984.282064] env[63355]: DEBUG nova.network.neutron [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 984.335239] env[63355]: INFO nova.compute.manager [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] instance snapshotting [ 984.336037] env[63355]: DEBUG nova.objects.instance [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'flavor' on Instance uuid 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.424997] env[63355]: DEBUG nova.network.neutron [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Updating instance_info_cache with network_info: [{"id": "677de229-5411-464b-9680-0080e5794772", "address": "fa:16:3e:03:96:b7", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap677de229-54", "ovs_interfaceid": "677de229-5411-464b-9680-0080e5794772", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.437104] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349940, 'name': PowerOffVM_Task, 'duration_secs': 0.383929} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.437422] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.438257] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8483ea6a-8cf7-444b-89db-c2c03c0375cf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.461039] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce09ca2-bbee-4097-af49-eae8f690fef7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.497490] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521673c4-b48b-cad0-6cc8-8f1941865d18, 'name': SearchDatastore_Task, 'duration_secs': 0.009595} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.498979] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e3540b6-6957-4c73-af55-8f46fb39d4b6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.505037] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 984.505037] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528a1917-f650-f877-ec9b-4d54adc9f3ee" [ 984.505037] env[63355]: _type = "Task" [ 984.505037] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.512373] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528a1917-f650-f877-ec9b-4d54adc9f3ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.580854] env[63355]: DEBUG nova.compute.manager [req-1c944f2e-9aa2-4c0f-b595-3aff6642d5c4 req-3829fa62-060c-47cf-95fe-86550e1319ab service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Received event network-vif-deleted-eb47f210-8e1e-4310-b314-33a489cf4325 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.581076] env[63355]: INFO nova.compute.manager [req-1c944f2e-9aa2-4c0f-b595-3aff6642d5c4 req-3829fa62-060c-47cf-95fe-86550e1319ab service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Neutron deleted interface eb47f210-8e1e-4310-b314-33a489cf4325; detaching it from the instance and deleting it from the info cache [ 984.581258] env[63355]: DEBUG nova.network.neutron [req-1c944f2e-9aa2-4c0f-b595-3aff6642d5c4 req-3829fa62-060c-47cf-95fe-86550e1319ab service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.652887] env[63355]: DEBUG nova.compute.utils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.656094] env[63355]: DEBUG nova.compute.manager [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 984.656312] env[63355]: DEBUG nova.network.neutron [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 984.692571] env[63355]: DEBUG nova.policy [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77243643ea724b72858a8682a2a054a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d47b671ea9c429391cbdae7e24adadf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.842585] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045f08e2-88ee-42c2-8216-8997f9911707 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.862273] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0472fa8d-51db-4870-8555-4a290a78bca6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.927453] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "refresh_cache-edb14fe7-d444-4fef-8c5d-d5616676e326" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.927787] env[63355]: DEBUG nova.compute.manager [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Instance network_info: |[{"id": "677de229-5411-464b-9680-0080e5794772", "address": "fa:16:3e:03:96:b7", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap677de229-54", "ovs_interfaceid": "677de229-5411-464b-9680-0080e5794772", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 984.928716] env[63355]: DEBUG nova.network.neutron [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Successfully created port: aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.930712] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:96:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '677de229-5411-464b-9680-0080e5794772', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.938187] env[63355]: DEBUG oslo.service.loopingcall [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.938997] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 984.939552] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe28512b-84c5-432f-a451-079597111cf9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.963531] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.963531] env[63355]: value = "task-1349941" [ 984.963531] env[63355]: _type = "Task" [ 984.963531] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.968774] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 984.969113] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d00c424e-266e-4294-9a5d-254dd5a15453 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.979886] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349941, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.981401] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 984.981401] env[63355]: value = "task-1349942" [ 984.981401] env[63355]: _type = "Task" [ 984.981401] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.993253] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349942, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.014822] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528a1917-f650-f877-ec9b-4d54adc9f3ee, 'name': SearchDatastore_Task, 'duration_secs': 0.041121} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.015126] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.015418] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] fd55ddfc-f376-4e15-961e-6826f31a7890/fd55ddfc-f376-4e15-961e-6826f31a7890.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 985.015703] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba8b4217-09ef-433c-bce7-3628fc281ebb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.021931] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 985.021931] env[63355]: value = "task-1349943" [ 985.021931] env[63355]: _type = "Task" [ 985.021931] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.030431] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.059832] env[63355]: DEBUG nova.network.neutron [-] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.084519] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5028a170-e4cf-4688-a549-4a78afc2e43b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.093399] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f053dd-7b44-4d23-ae94-b9e7f017cf2b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.122974] env[63355]: DEBUG nova.compute.manager [req-1c944f2e-9aa2-4c0f-b595-3aff6642d5c4 req-3829fa62-060c-47cf-95fe-86550e1319ab service nova] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Detach interface failed, port_id=eb47f210-8e1e-4310-b314-33a489cf4325, reason: Instance c630c5fe-6907-4952-9807-6e59bd1cc9e1 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 985.157078] env[63355]: DEBUG nova.compute.manager [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 985.373535] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 985.376723] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4befec4d-96df-487a-a460-72f1cb486b69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.386563] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 985.386563] env[63355]: value = "task-1349944" [ 985.386563] env[63355]: _type = "Task" [ 985.386563] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.399137] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349944, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.478740] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349941, 'name': CreateVM_Task, 'duration_secs': 0.338232} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.478938] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.479701] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.479905] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.480271] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.480555] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5222c9f5-150e-41f9-8b9a-b0caf16219d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.483905] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa50170-1e76-4589-b9cc-6e286d5ac642 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.498518] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd30f38d-7c5a-44c4-a3dc-1f03a8189849 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.501736] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349942, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.502020] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 985.502020] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5293f875-c931-fbfc-7d94-bc2321259329" [ 985.502020] env[63355]: _type = "Task" [ 985.502020] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.532594] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7413f3bb-eb2d-4378-be8c-0c457eeb81d5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.540014] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5293f875-c931-fbfc-7d94-bc2321259329, 'name': SearchDatastore_Task, 'duration_secs': 0.011346} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.541145] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.541483] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.541853] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.542150] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.542426] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.547381] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c08cbb79-2383-412a-a4ab-8b1b2e8eb61f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.549731] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349943, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.551199] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b690ab-ca33-478f-b39b-ce802754e446 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.568483] env[63355]: INFO nova.compute.manager [-] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Took 1.29 seconds to deallocate network for instance. [ 985.569019] env[63355]: DEBUG nova.compute.provider_tree [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.572643] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.572830] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 985.574071] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d564bbb3-fbee-45a6-a818-050f527a0ef3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.583700] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 985.583700] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527167fa-ea3b-f50f-1bed-297683557927" [ 985.583700] env[63355]: _type = "Task" [ 985.583700] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.592885] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527167fa-ea3b-f50f-1bed-297683557927, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.896048] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349944, 'name': CreateSnapshot_Task, 'duration_secs': 0.489677} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.896048] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 985.896369] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d51980f-bc83-41e7-97d9-ac81ececabac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.992661] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349942, 'name': CreateSnapshot_Task, 'duration_secs': 0.630667} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.993066] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 985.993678] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9161236-c8fd-4185-985c-2305c6940b6f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.038888] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543683} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.039157] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] fd55ddfc-f376-4e15-961e-6826f31a7890/fd55ddfc-f376-4e15-961e-6826f31a7890.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 986.039433] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.039710] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d67ca508-d328-4ae5-afad-2f1440a0acc6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.046417] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 986.046417] env[63355]: value = "task-1349945" [ 986.046417] env[63355]: _type = "Task" [ 986.046417] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.053496] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349945, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.074894] env[63355]: DEBUG nova.scheduler.client.report [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.081349] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.092629] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527167fa-ea3b-f50f-1bed-297683557927, 'name': SearchDatastore_Task, 'duration_secs': 0.018704} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.093969] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73a7e2d1-266b-4bf3-9898-108b014fc008 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.099446] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 986.099446] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c663dd-b023-0095-99ff-b322a7b2691c" [ 986.099446] env[63355]: _type = "Task" [ 986.099446] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.108190] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c663dd-b023-0095-99ff-b322a7b2691c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.171543] env[63355]: DEBUG nova.compute.manager [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 986.192686] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.192934] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.193108] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.193336] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.193493] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.193643] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.193851] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.194035] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.194241] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.194415] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.194595] env[63355]: DEBUG nova.virt.hardware [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.195437] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254df50d-93f8-46e0-bd49-94ee8f7a1400 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.204816] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69641fd7-3f78-4a70-ac9a-61c2efef37d0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.412682] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 986.413009] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f8fce08c-4740-401f-95b1-45947d15cc9c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.421749] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 986.421749] env[63355]: value = "task-1349946" [ 986.421749] env[63355]: _type = "Task" [ 986.421749] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.431041] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349946, 'name': CloneVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.510186] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 986.510651] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1fd9a873-1d83-40a3-9fc6-572ee7d5a33b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.518448] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 986.518448] env[63355]: value = "task-1349947" [ 986.518448] env[63355]: _type = "Task" [ 986.518448] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.526724] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349947, 'name': CloneVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.555199] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349945, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066145} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.555451] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 986.556750] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca3acfa-fb7e-43fe-b63e-665611c79699 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.579648] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] fd55ddfc-f376-4e15-961e-6826f31a7890/fd55ddfc-f376-4e15-961e-6826f31a7890.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.580464] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.581087] env[63355]: DEBUG nova.compute.manager [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 986.583905] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38922f91-6101-4278-82ba-5c4fe22fc1bb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.598906] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.795s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.599217] env[63355]: DEBUG nova.objects.instance [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lazy-loading 'resources' on Instance uuid 248ec2f2-3a41-444a-b550-71c5438a45f7 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.607580] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 986.607580] env[63355]: value = "task-1349948" [ 986.607580] env[63355]: _type = "Task" [ 986.607580] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.608808] env[63355]: DEBUG nova.compute.manager [req-ccd9ef6d-f28f-4f2c-96d9-25b87c455ef3 req-962a3878-646b-445f-85db-473f8ef4e0aa service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received event network-vif-plugged-aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.609113] env[63355]: DEBUG oslo_concurrency.lockutils [req-ccd9ef6d-f28f-4f2c-96d9-25b87c455ef3 req-962a3878-646b-445f-85db-473f8ef4e0aa service nova] Acquiring lock "da3cb83c-6368-49c0-9b11-0498221e3c0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.609752] env[63355]: DEBUG oslo_concurrency.lockutils [req-ccd9ef6d-f28f-4f2c-96d9-25b87c455ef3 req-962a3878-646b-445f-85db-473f8ef4e0aa service nova] Lock "da3cb83c-6368-49c0-9b11-0498221e3c0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.609752] env[63355]: DEBUG oslo_concurrency.lockutils [req-ccd9ef6d-f28f-4f2c-96d9-25b87c455ef3 req-962a3878-646b-445f-85db-473f8ef4e0aa service nova] Lock "da3cb83c-6368-49c0-9b11-0498221e3c0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.609870] env[63355]: DEBUG nova.compute.manager [req-ccd9ef6d-f28f-4f2c-96d9-25b87c455ef3 req-962a3878-646b-445f-85db-473f8ef4e0aa service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] No waiting events found dispatching network-vif-plugged-aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 986.610422] env[63355]: WARNING nova.compute.manager [req-ccd9ef6d-f28f-4f2c-96d9-25b87c455ef3 req-962a3878-646b-445f-85db-473f8ef4e0aa service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received unexpected event network-vif-plugged-aa5a40b6-5d12-4d62-88d2-32a26a65a241 for instance with vm_state building and task_state spawning. [ 986.620518] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c663dd-b023-0095-99ff-b322a7b2691c, 'name': SearchDatastore_Task, 'duration_secs': 0.0111} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.621318] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.621635] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] edb14fe7-d444-4fef-8c5d-d5616676e326/edb14fe7-d444-4fef-8c5d-d5616676e326.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 986.622206] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ef6fbf0-9b08-4437-81da-0b46dfe92f37 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.628907] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.633259] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 986.633259] env[63355]: value = "task-1349949" [ 986.633259] env[63355]: _type = "Task" [ 986.633259] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.641752] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349949, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.822352] env[63355]: DEBUG nova.network.neutron [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Successfully updated port: aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.933611] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349946, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.030754] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349947, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.100314] env[63355]: DEBUG nova.compute.utils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.101830] env[63355]: DEBUG nova.compute.manager [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.102036] env[63355]: DEBUG nova.network.neutron [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 987.122431] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.146955] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349949, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.149046] env[63355]: DEBUG nova.policy [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24c7a31ced4042eb942ab42514eaa96b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '625f9b0e39914a1c84736cc3effdc11b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 987.295238] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fe59ba-24b4-4a1e-993b-19bbe4378658 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.303722] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d8b642-7f91-40fb-a18f-fddd0d58dcfa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.334338] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.334438] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.334648] env[63355]: DEBUG nova.network.neutron [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 987.337032] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5c2bdd-5bf0-4e98-b6ad-4eb8c4234ecb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.351121] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1343c8-ee8c-4b4c-96b6-ba21ee76d47a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.367102] env[63355]: DEBUG nova.compute.provider_tree [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.435981] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349946, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.532799] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349947, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.558546] env[63355]: DEBUG nova.network.neutron [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Successfully created port: f7b42e7c-9480-4b58-a4de-69fb8b6c76ce {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.608701] env[63355]: DEBUG nova.compute.manager [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 987.629193] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349948, 'name': ReconfigVM_Task, 'duration_secs': 0.952962} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.631908] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Reconfigured VM instance instance-00000055 to attach disk [datastore2] fd55ddfc-f376-4e15-961e-6826f31a7890/fd55ddfc-f376-4e15-961e-6826f31a7890.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.632739] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1fa5f71d-5deb-4091-a936-bdb9d117ad95 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.645475] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 987.645475] env[63355]: value = "task-1349950" [ 987.645475] env[63355]: _type = "Task" [ 987.645475] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.654423] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349949, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.753066} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.657979] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] edb14fe7-d444-4fef-8c5d-d5616676e326/edb14fe7-d444-4fef-8c5d-d5616676e326.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 987.658244] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.658526] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57cd8574-4e06-4554-8b91-830dd5c2f5fa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.665832] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349950, 'name': Rename_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.668129] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 987.668129] env[63355]: value = "task-1349951" [ 987.668129] env[63355]: _type = "Task" [ 987.668129] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.677540] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349951, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.871645] env[63355]: DEBUG nova.scheduler.client.report [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.878408] env[63355]: DEBUG nova.network.neutron [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 987.940424] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349946, 'name': CloneVM_Task, 'duration_secs': 1.372404} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.940711] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Created linked-clone VM from snapshot [ 987.941533] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842083bd-c7bb-4162-b0bb-b218af2f4026 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.953245] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Uploading image dfe52617-1814-4a03-b7f0-4f5b8cfa9c9c {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 987.975311] env[63355]: DEBUG oslo_vmware.rw_handles [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 987.975311] env[63355]: value = "vm-287726" [ 987.975311] env[63355]: _type = "VirtualMachine" [ 987.975311] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 987.975611] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-35a8b0d3-53c9-4d52-82a2-21df936bcc34 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.983148] env[63355]: DEBUG oslo_vmware.rw_handles [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lease: (returnval){ [ 987.983148] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52679766-5638-8c95-45a6-e5b1464817c0" [ 987.983148] env[63355]: _type = "HttpNfcLease" [ 987.983148] env[63355]: } obtained for exporting VM: (result){ [ 987.983148] env[63355]: value = "vm-287726" [ 987.983148] env[63355]: _type = "VirtualMachine" [ 987.983148] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 987.983458] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the lease: (returnval){ [ 987.983458] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52679766-5638-8c95-45a6-e5b1464817c0" [ 987.983458] env[63355]: _type = "HttpNfcLease" [ 987.983458] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 987.989878] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 987.989878] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52679766-5638-8c95-45a6-e5b1464817c0" [ 987.989878] env[63355]: _type = "HttpNfcLease" [ 987.989878] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 988.029256] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349947, 'name': CloneVM_Task, 'duration_secs': 1.330702} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.029612] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Created linked-clone VM from snapshot [ 988.030428] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b74aa0d-8f58-4694-a26f-5bd63ea29ca7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.037999] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Uploading image c58460ed-ef70-483a-87f9-6152e2a0d383 {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 988.067133] env[63355]: DEBUG oslo_vmware.rw_handles [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 988.067133] env[63355]: value = "vm-287727" [ 988.067133] env[63355]: _type = "VirtualMachine" [ 988.067133] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 988.067453] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e84978ac-a12e-44a3-8973-8fd01b6dab94 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.073433] env[63355]: DEBUG oslo_vmware.rw_handles [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lease: (returnval){ [ 988.073433] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5258eae1-b94e-77e2-fdf6-d1bc57277717" [ 988.073433] env[63355]: _type = "HttpNfcLease" [ 988.073433] env[63355]: } obtained for exporting VM: (result){ [ 988.073433] env[63355]: value = "vm-287727" [ 988.073433] env[63355]: _type = "VirtualMachine" [ 988.073433] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 988.073707] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the lease: (returnval){ [ 988.073707] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5258eae1-b94e-77e2-fdf6-d1bc57277717" [ 988.073707] env[63355]: _type = "HttpNfcLease" [ 988.073707] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 988.079686] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 988.079686] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5258eae1-b94e-77e2-fdf6-d1bc57277717" [ 988.079686] env[63355]: _type = "HttpNfcLease" [ 988.079686] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 988.163380] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349950, 'name': Rename_Task, 'duration_secs': 0.12917} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.164308] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 988.166569] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b71d34f-3444-41bf-804d-597857674797 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.174243] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 988.174243] env[63355]: value = "task-1349954" [ 988.174243] env[63355]: _type = "Task" [ 988.174243] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.181241] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349951, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067877} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.181754] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.182625] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e5945e5-7abc-42a5-9bdb-726e8a78701d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.188782] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.208017] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] edb14fe7-d444-4fef-8c5d-d5616676e326/edb14fe7-d444-4fef-8c5d-d5616676e326.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.208983] env[63355]: DEBUG nova.network.neutron [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [{"id": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "address": "fa:16:3e:08:9e:2a", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5a40b6-5d", "ovs_interfaceid": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.210333] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f9b59e6-6acf-45a2-b2ba-954e4c45ba35 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.226301] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.226606] env[63355]: DEBUG nova.compute.manager [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Instance network_info: |[{"id": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "address": "fa:16:3e:08:9e:2a", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5a40b6-5d", "ovs_interfaceid": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 988.227295] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:9e:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e41070eb-3ac1-4ca9-a3d0-fd65893a97de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa5a40b6-5d12-4d62-88d2-32a26a65a241', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 988.234470] env[63355]: DEBUG oslo.service.loopingcall [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.235321] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 988.235634] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-494cde3b-ef26-499c-bb40-991a87b7a0bd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.251613] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 988.251613] env[63355]: value = "task-1349955" [ 988.251613] env[63355]: _type = "Task" [ 988.251613] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.256176] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 988.256176] env[63355]: value = "task-1349956" [ 988.256176] env[63355]: _type = "Task" [ 988.256176] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.263231] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349955, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.267805] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349956, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.375887] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.777s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.378722] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.178s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.379030] env[63355]: DEBUG nova.objects.instance [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lazy-loading 'resources' on Instance uuid cd9de9b8-ad42-4dbb-b435-927738a55f3f {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.402028] env[63355]: INFO nova.scheduler.client.report [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleted allocations for instance 248ec2f2-3a41-444a-b550-71c5438a45f7 [ 988.492365] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 988.492365] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52679766-5638-8c95-45a6-e5b1464817c0" [ 988.492365] env[63355]: _type = "HttpNfcLease" [ 988.492365] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 988.492755] env[63355]: DEBUG oslo_vmware.rw_handles [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 988.492755] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52679766-5638-8c95-45a6-e5b1464817c0" [ 988.492755] env[63355]: _type = "HttpNfcLease" [ 988.492755] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 988.493580] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a81b3d-5fd4-4676-8c50-a4b85ce006d6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.501243] env[63355]: DEBUG oslo_vmware.rw_handles [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524bce97-3be4-2e76-81b6-8c4ca3306e6f/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 988.501512] env[63355]: DEBUG oslo_vmware.rw_handles [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524bce97-3be4-2e76-81b6-8c4ca3306e6f/disk-0.vmdk for reading. {{(pid=63355) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 988.581776] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 988.581776] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5258eae1-b94e-77e2-fdf6-d1bc57277717" [ 988.581776] env[63355]: _type = "HttpNfcLease" [ 988.581776] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 988.582102] env[63355]: DEBUG oslo_vmware.rw_handles [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 988.582102] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5258eae1-b94e-77e2-fdf6-d1bc57277717" [ 988.582102] env[63355]: _type = "HttpNfcLease" [ 988.582102] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 988.582814] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274a4afb-b5b6-4709-b98d-de2c4af3118f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.591628] env[63355]: DEBUG oslo_vmware.rw_handles [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52399965-bfc6-b4c7-b8e9-12168bc4480e/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 988.591833] env[63355]: DEBUG oslo_vmware.rw_handles [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52399965-bfc6-b4c7-b8e9-12168bc4480e/disk-0.vmdk for reading. {{(pid=63355) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 988.595305] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-996cca48-8d9b-4a90-9b7d-66c3405f9feb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.655923] env[63355]: DEBUG nova.compute.manager [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 988.687857] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.688143] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.688299] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.688500] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.688684] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.688842] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.689068] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.689261] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.689423] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.689622] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.689762] env[63355]: DEBUG nova.virt.hardware [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.690633] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d807ef-83d3-4804-87c0-2601e39befc4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.696406] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349954, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.704899] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e067c839-cc09-43e6-b46e-51f0d2509a5e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.720459] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-012656a0-3405-45fd-9e96-a58c14535603 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.724400] env[63355]: DEBUG nova.compute.manager [req-2451b9a7-f9a9-414e-9cfb-55ea5b2e1ba3 req-7b7b026a-85ce-4b0f-8f72-347d0dcdccbd service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received event network-changed-aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.724592] env[63355]: DEBUG nova.compute.manager [req-2451b9a7-f9a9-414e-9cfb-55ea5b2e1ba3 req-7b7b026a-85ce-4b0f-8f72-347d0dcdccbd service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing instance network info cache due to event network-changed-aa5a40b6-5d12-4d62-88d2-32a26a65a241. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 988.724824] env[63355]: DEBUG oslo_concurrency.lockutils [req-2451b9a7-f9a9-414e-9cfb-55ea5b2e1ba3 req-7b7b026a-85ce-4b0f-8f72-347d0dcdccbd service nova] Acquiring lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.725090] env[63355]: DEBUG oslo_concurrency.lockutils [req-2451b9a7-f9a9-414e-9cfb-55ea5b2e1ba3 req-7b7b026a-85ce-4b0f-8f72-347d0dcdccbd service nova] Acquired lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.726018] env[63355]: DEBUG nova.network.neutron [req-2451b9a7-f9a9-414e-9cfb-55ea5b2e1ba3 req-7b7b026a-85ce-4b0f-8f72-347d0dcdccbd service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing network info cache for port aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.762315] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349955, 'name': ReconfigVM_Task, 'duration_secs': 0.268735} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.763987] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Reconfigured VM instance instance-00000056 to attach disk [datastore2] edb14fe7-d444-4fef-8c5d-d5616676e326/edb14fe7-d444-4fef-8c5d-d5616676e326.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.763987] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37522976-65ce-44ee-b6bc-6a34650c355b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.767832] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349956, 'name': CreateVM_Task, 'duration_secs': 0.367318} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.768266] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.768904] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.769070] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.769381] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.769853] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a287aeea-44fe-4e0b-916c-50be24320bb9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.772589] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 988.772589] env[63355]: value = "task-1349957" [ 988.772589] env[63355]: _type = "Task" [ 988.772589] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.776324] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 988.776324] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d8f3db-e6c6-d94a-8bcb-ffde59276092" [ 988.776324] env[63355]: _type = "Task" [ 988.776324] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.782616] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349957, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.787412] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d8f3db-e6c6-d94a-8bcb-ffde59276092, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.881833] env[63355]: DEBUG nova.objects.instance [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lazy-loading 'numa_topology' on Instance uuid cd9de9b8-ad42-4dbb-b435-927738a55f3f {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.912484] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d63a3067-55ad-4c6b-99f8-3b249110cebf tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "248ec2f2-3a41-444a-b550-71c5438a45f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.637s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.191128] env[63355]: DEBUG oslo_vmware.api [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1349954, 'name': PowerOnVM_Task, 'duration_secs': 0.540159} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.191523] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 989.193044] env[63355]: INFO nova.compute.manager [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Took 9.46 seconds to spawn the instance on the hypervisor. [ 989.193044] env[63355]: DEBUG nova.compute.manager [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 989.193044] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47411fbd-851d-4043-b90c-1cb9c3c69c6a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.290879] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349957, 'name': Rename_Task, 'duration_secs': 0.162202} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.295387] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 989.295734] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d8f3db-e6c6-d94a-8bcb-ffde59276092, 'name': SearchDatastore_Task, 'duration_secs': 0.026895} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.296978] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ab89237-3072-4259-a317-95e014a88436 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.298874] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.299157] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.300594] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.300806] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.301451] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.301769] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9819f57a-ee80-4ff7-8763-d255e5a2fa26 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.311822] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 989.311822] env[63355]: value = "task-1349958" [ 989.311822] env[63355]: _type = "Task" [ 989.311822] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.319287] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.319563] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.320859] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5553666d-199b-4e6d-8044-d1926a0c977c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.328829] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349958, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.336222] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 989.336222] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f9f0d1-5411-4449-5773-dc1c331c5e54" [ 989.336222] env[63355]: _type = "Task" [ 989.336222] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.346041] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f9f0d1-5411-4449-5773-dc1c331c5e54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.388396] env[63355]: DEBUG nova.objects.base [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 989.460741] env[63355]: DEBUG nova.network.neutron [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Successfully updated port: f7b42e7c-9480-4b58-a4de-69fb8b6c76ce {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.503065] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.503360] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.503678] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.503850] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.504045] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.509384] env[63355]: INFO nova.compute.manager [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Terminating instance [ 989.512264] env[63355]: DEBUG nova.compute.manager [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 989.512488] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 989.513770] env[63355]: DEBUG nova.network.neutron [req-2451b9a7-f9a9-414e-9cfb-55ea5b2e1ba3 req-7b7b026a-85ce-4b0f-8f72-347d0dcdccbd service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updated VIF entry in instance network info cache for port aa5a40b6-5d12-4d62-88d2-32a26a65a241. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 989.514241] env[63355]: DEBUG nova.network.neutron [req-2451b9a7-f9a9-414e-9cfb-55ea5b2e1ba3 req-7b7b026a-85ce-4b0f-8f72-347d0dcdccbd service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [{"id": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "address": "fa:16:3e:08:9e:2a", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5a40b6-5d", "ovs_interfaceid": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.516322] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5931d0-4fa2-4509-8d25-bf10e218f697 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.527525] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 989.528786] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db7fed9e-96c4-4507-bb7c-42be560265c5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.537264] env[63355]: DEBUG oslo_vmware.api [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 989.537264] env[63355]: value = "task-1349959" [ 989.537264] env[63355]: _type = "Task" [ 989.537264] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.553354] env[63355]: DEBUG oslo_vmware.api [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349959, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.614246] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167766b7-c77f-4fa9-b9d6-32ce797eaf35 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.622377] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a9a771-bec6-4dea-883b-d107235a3c10 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.656579] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f1dcaf-319b-4264-b022-9d1cbb707cd7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.665895] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43caafd4-507e-47c0-8270-51e0083199d0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.680998] env[63355]: DEBUG nova.compute.provider_tree [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.718556] env[63355]: INFO nova.compute.manager [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Took 17.46 seconds to build instance. [ 989.822661] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349958, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.847985] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f9f0d1-5411-4449-5773-dc1c331c5e54, 'name': SearchDatastore_Task, 'duration_secs': 0.01316} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.849311] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92a3f4f1-5d2a-4a72-8308-9047ccdb0bc1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.855254] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 989.855254] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e7b7c4-7bfd-146f-8c94-aa820ae03dcc" [ 989.855254] env[63355]: _type = "Task" [ 989.855254] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.865294] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e7b7c4-7bfd-146f-8c94-aa820ae03dcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.965261] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquiring lock "refresh_cache-9ee3001b-8bf0-43ab-996a-a68dad57d8e6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.965261] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquired lock "refresh_cache-9ee3001b-8bf0-43ab-996a-a68dad57d8e6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.965261] env[63355]: DEBUG nova.network.neutron [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.021830] env[63355]: DEBUG oslo_concurrency.lockutils [req-2451b9a7-f9a9-414e-9cfb-55ea5b2e1ba3 req-7b7b026a-85ce-4b0f-8f72-347d0dcdccbd service nova] Releasing lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.056635] env[63355]: DEBUG oslo_vmware.api [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349959, 'name': PowerOffVM_Task, 'duration_secs': 0.229907} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.057501] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 990.058027] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 990.058854] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-00cdbf7f-aeef-4b75-8292-bf7bf1c13b3b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.140701] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 990.141517] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 990.141517] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleting the datastore file [datastore2] f5e62ce1-40b7-4648-a4a6-068ff06eaf9b {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.141775] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-181cb200-5786-4287-a8ef-c9d147d31e8f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.149089] env[63355]: DEBUG oslo_vmware.api [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 990.149089] env[63355]: value = "task-1349961" [ 990.149089] env[63355]: _type = "Task" [ 990.149089] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.157767] env[63355]: DEBUG oslo_vmware.api [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349961, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.185207] env[63355]: DEBUG nova.scheduler.client.report [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.221594] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1252fdd3-13df-4f88-8c4b-3c58a0c6c768 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.977s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.323397] env[63355]: DEBUG oslo_vmware.api [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1349958, 'name': PowerOnVM_Task, 'duration_secs': 0.638737} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.324018] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 990.324338] env[63355]: INFO nova.compute.manager [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Took 8.22 seconds to spawn the instance on the hypervisor. [ 990.324666] env[63355]: DEBUG nova.compute.manager [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.325539] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a8bbe9e-b233-47de-a004-47237dc9dfab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.371019] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e7b7c4-7bfd-146f-8c94-aa820ae03dcc, 'name': SearchDatastore_Task, 'duration_secs': 0.012885} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.371655] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.372075] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] da3cb83c-6368-49c0-9b11-0498221e3c0f/da3cb83c-6368-49c0-9b11-0498221e3c0f.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 990.372426] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d7a058d-6be5-43a9-aa13-e7fea786deee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.379287] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 990.379287] env[63355]: value = "task-1349962" [ 990.379287] env[63355]: _type = "Task" [ 990.379287] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.389186] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349962, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.501349] env[63355]: DEBUG nova.network.neutron [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 990.655735] env[63355]: DEBUG nova.network.neutron [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Updating instance_info_cache with network_info: [{"id": "f7b42e7c-9480-4b58-a4de-69fb8b6c76ce", "address": "fa:16:3e:c3:db:cf", "network": {"id": "a80daa41-54f9-415b-a3d7-1dbf1800cc4a", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1679031355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "625f9b0e39914a1c84736cc3effdc11b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7b42e7c-94", "ovs_interfaceid": "f7b42e7c-9480-4b58-a4de-69fb8b6c76ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.663698] env[63355]: DEBUG oslo_vmware.api [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1349961, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26164} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.664175] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.664384] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.664505] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.664645] env[63355]: INFO nova.compute.manager [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Took 1.15 seconds to destroy the instance on the hypervisor. [ 990.664867] env[63355]: DEBUG oslo.service.loopingcall [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.665213] env[63355]: DEBUG nova.compute.manager [-] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.665306] env[63355]: DEBUG nova.network.neutron [-] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 990.690669] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.312s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.694018] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.585s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.849501] env[63355]: INFO nova.compute.manager [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Took 17.97 seconds to build instance. [ 990.892359] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349962, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.903755] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e4d89e-54c9-4699-9ca2-7f9ce34a7c32 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.911706] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db67a6f5-de6e-4f45-8442-06ee928df4e3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.949864] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d533b01f-789c-4a48-aaca-ab3c11ad3421 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.958771] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5c5dd8-70a4-4bfc-a8ed-859a86f7ed7a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.975961] env[63355]: DEBUG nova.compute.provider_tree [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.042060] env[63355]: DEBUG nova.compute.manager [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Received event network-vif-plugged-f7b42e7c-9480-4b58-a4de-69fb8b6c76ce {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.042451] env[63355]: DEBUG oslo_concurrency.lockutils [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] Acquiring lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.043037] env[63355]: DEBUG oslo_concurrency.lockutils [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] Lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.043250] env[63355]: DEBUG oslo_concurrency.lockutils [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] Lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.043479] env[63355]: DEBUG nova.compute.manager [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] No waiting events found dispatching network-vif-plugged-f7b42e7c-9480-4b58-a4de-69fb8b6c76ce {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 991.043603] env[63355]: WARNING nova.compute.manager [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Received unexpected event network-vif-plugged-f7b42e7c-9480-4b58-a4de-69fb8b6c76ce for instance with vm_state building and task_state spawning. [ 991.043800] env[63355]: DEBUG nova.compute.manager [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Received event network-changed-f7b42e7c-9480-4b58-a4de-69fb8b6c76ce {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.043985] env[63355]: DEBUG nova.compute.manager [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Refreshing instance network info cache due to event network-changed-f7b42e7c-9480-4b58-a4de-69fb8b6c76ce. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.044194] env[63355]: DEBUG oslo_concurrency.lockutils [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] Acquiring lock "refresh_cache-9ee3001b-8bf0-43ab-996a-a68dad57d8e6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.160141] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Releasing lock "refresh_cache-9ee3001b-8bf0-43ab-996a-a68dad57d8e6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.160539] env[63355]: DEBUG nova.compute.manager [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Instance network_info: |[{"id": "f7b42e7c-9480-4b58-a4de-69fb8b6c76ce", "address": "fa:16:3e:c3:db:cf", "network": {"id": "a80daa41-54f9-415b-a3d7-1dbf1800cc4a", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1679031355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "625f9b0e39914a1c84736cc3effdc11b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7b42e7c-94", "ovs_interfaceid": "f7b42e7c-9480-4b58-a4de-69fb8b6c76ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 991.160895] env[63355]: DEBUG oslo_concurrency.lockutils [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] Acquired lock "refresh_cache-9ee3001b-8bf0-43ab-996a-a68dad57d8e6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.161123] env[63355]: DEBUG nova.network.neutron [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Refreshing network info cache for port f7b42e7c-9480-4b58-a4de-69fb8b6c76ce {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.162622] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:db:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f7b42e7c-9480-4b58-a4de-69fb8b6c76ce', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.171238] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Creating folder: Project (625f9b0e39914a1c84736cc3effdc11b). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 991.174930] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-99e5daf8-3681-49b8-b5a3-e245dcdcf551 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.188945] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Created folder: Project (625f9b0e39914a1c84736cc3effdc11b) in parent group-v287607. [ 991.189192] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Creating folder: Instances. Parent ref: group-v287729. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 991.189457] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1fd1a54c-b4a4-4977-a764-da14b847e14b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.203026] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Created folder: Instances in parent group-v287729. [ 991.203026] env[63355]: DEBUG oslo.service.loopingcall [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.203650] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.203650] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-659c489b-40c2-4cdc-8a08-9f32fb2f39bb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.223965] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c3d21582-4443-4a2e-93c5-d668dcb6e37d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 38.311s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.226558] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 14.573s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.226908] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.227234] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.227538] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.230168] env[63355]: DEBUG nova.compute.manager [req-5eef39a5-1ef2-4b31-b500-479f9bc0159a req-f27df63b-0a44-4b82-94ce-cb94d9d29f79 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Received event network-vif-deleted-ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.230443] env[63355]: INFO nova.compute.manager [req-5eef39a5-1ef2-4b31-b500-479f9bc0159a req-f27df63b-0a44-4b82-94ce-cb94d9d29f79 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Neutron deleted interface ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5; detaching it from the instance and deleting it from the info cache [ 991.231621] env[63355]: DEBUG nova.network.neutron [req-5eef39a5-1ef2-4b31-b500-479f9bc0159a req-f27df63b-0a44-4b82-94ce-cb94d9d29f79 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.232296] env[63355]: INFO nova.compute.manager [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Terminating instance [ 991.235323] env[63355]: DEBUG nova.compute.manager [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 991.235575] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 991.236218] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94b79db3-7172-4e56-834d-d150c5f6e2b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.240281] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.240281] env[63355]: value = "task-1349965" [ 991.240281] env[63355]: _type = "Task" [ 991.240281] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.252265] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc327c4a-2fa1-4db5-a826-a042a5249a85 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.268812] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349965, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.290042] env[63355]: WARNING nova.virt.vmwareapi.vmops [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cd9de9b8-ad42-4dbb-b435-927738a55f3f could not be found. [ 991.290199] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 991.290498] env[63355]: INFO nova.compute.manager [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 991.290931] env[63355]: DEBUG oslo.service.loopingcall [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.294332] env[63355]: DEBUG nova.compute.manager [-] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 991.294556] env[63355]: DEBUG nova.network.neutron [-] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 991.351519] env[63355]: DEBUG oslo_concurrency.lockutils [None req-40ada902-d139-47f5-b7fe-2a6d561aa331 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "edb14fe7-d444-4fef-8c5d-d5616676e326" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.489s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.394876] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349962, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.66789} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.397811] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] da3cb83c-6368-49c0-9b11-0498221e3c0f/da3cb83c-6368-49c0-9b11-0498221e3c0f.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.397811] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.397811] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9bf0f05b-e280-487e-871a-bc5c25c2d477 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.402945] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 991.402945] env[63355]: value = "task-1349966" [ 991.402945] env[63355]: _type = "Task" [ 991.402945] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.412901] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349966, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.454386] env[63355]: DEBUG nova.network.neutron [-] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.474506] env[63355]: DEBUG nova.network.neutron [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Updated VIF entry in instance network info cache for port f7b42e7c-9480-4b58-a4de-69fb8b6c76ce. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.474893] env[63355]: DEBUG nova.network.neutron [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Updating instance_info_cache with network_info: [{"id": "f7b42e7c-9480-4b58-a4de-69fb8b6c76ce", "address": "fa:16:3e:c3:db:cf", "network": {"id": "a80daa41-54f9-415b-a3d7-1dbf1800cc4a", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1679031355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "625f9b0e39914a1c84736cc3effdc11b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7b42e7c-94", "ovs_interfaceid": "f7b42e7c-9480-4b58-a4de-69fb8b6c76ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.479620] env[63355]: DEBUG nova.scheduler.client.report [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.737274] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f8d7180-05f3-46f3-884a-0e716631d9b7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.749337] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6916a1-b3bc-4cb2-8ef1-96952869fbea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.763111] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349965, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.784778] env[63355]: DEBUG nova.compute.manager [req-5eef39a5-1ef2-4b31-b500-479f9bc0159a req-f27df63b-0a44-4b82-94ce-cb94d9d29f79 service nova] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Detach interface failed, port_id=ec8f312b-be42-4ae6-a1b4-ebc98e64b4f5, reason: Instance f5e62ce1-40b7-4648-a4a6-068ff06eaf9b could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 991.912620] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349966, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098494} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.912919] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.913904] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4477144f-bd6d-4c9d-8cbc-5b66bf8b34a5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.938283] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] da3cb83c-6368-49c0-9b11-0498221e3c0f/da3cb83c-6368-49c0-9b11-0498221e3c0f.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.938600] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab19d751-cc60-4c1b-9493-b472f2fe7b6f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.958511] env[63355]: INFO nova.compute.manager [-] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Took 1.29 seconds to deallocate network for instance. [ 991.958953] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 991.958953] env[63355]: value = "task-1349967" [ 991.958953] env[63355]: _type = "Task" [ 991.958953] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.971284] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349967, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.977888] env[63355]: DEBUG oslo_concurrency.lockutils [req-7a873b84-0980-495c-995d-916c3b524361 req-ef34bd06-02fd-4f34-ad78-dae01c178268 service nova] Releasing lock "refresh_cache-9ee3001b-8bf0-43ab-996a-a68dad57d8e6" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.985080] env[63355]: DEBUG oslo_concurrency.lockutils [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.291s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.985385] env[63355]: INFO nova.compute.manager [None req-62eab159-ce6a-4217-a421-51b4af8bc054 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Successfully reverted task state from rebuilding on failure for instance. [ 991.990994] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.918s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.991138] env[63355]: DEBUG nova.objects.instance [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lazy-loading 'resources' on Instance uuid 8e9ff9b7-636f-48df-9168-509d733278f8 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.121388] env[63355]: DEBUG nova.network.neutron [-] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.253373] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349965, 'name': CreateVM_Task, 'duration_secs': 0.598133} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.253761] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 992.254656] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.254872] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.255290] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 992.255592] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e73d24e3-305a-4281-b458-2051a602fdba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.260865] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for the task: (returnval){ [ 992.260865] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fdffdf-e575-c367-8117-64f23b23255c" [ 992.260865] env[63355]: _type = "Task" [ 992.260865] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.270084] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fdffdf-e575-c367-8117-64f23b23255c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.471333] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.471672] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.624060] env[63355]: INFO nova.compute.manager [-] [instance: cd9de9b8-ad42-4dbb-b435-927738a55f3f] Took 1.33 seconds to deallocate network for instance. [ 992.679685] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46afc066-1f8b-48b1-91a9-43b4d2d770de {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.688019] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a917128-f40a-4958-86f0-c56f0aadb1f2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.721230] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ff85ed-8b63-4463-a1ae-cd9d534411c7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.728848] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa43781-6382-4c10-aa99-837b65097a04 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.742651] env[63355]: DEBUG nova.compute.provider_tree [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.771223] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fdffdf-e575-c367-8117-64f23b23255c, 'name': SearchDatastore_Task, 'duration_secs': 0.018828} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.772022] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.772022] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 992.772022] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.772225] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.772375] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.772683] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2105f6ae-9b43-42f8-ac32-9f2cde05fb0f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.782285] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 992.782547] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 992.783296] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02e5efb5-2037-4969-864c-cc051dc9d779 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.788803] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for the task: (returnval){ [ 992.788803] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f2bcf9-531b-abdd-f7ad-68442ebebd41" [ 992.788803] env[63355]: _type = "Task" [ 992.788803] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.797309] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f2bcf9-531b-abdd-f7ad-68442ebebd41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.971159] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349967, 'name': ReconfigVM_Task, 'duration_secs': 0.718074} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.971446] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Reconfigured VM instance instance-00000057 to attach disk [datastore1] da3cb83c-6368-49c0-9b11-0498221e3c0f/da3cb83c-6368-49c0-9b11-0498221e3c0f.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.972106] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a247eb50-92a7-4343-9cd0-190acf38aa6a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.978920] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 992.978920] env[63355]: value = "task-1349968" [ 992.978920] env[63355]: _type = "Task" [ 992.978920] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.987420] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349968, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.246749] env[63355]: DEBUG nova.scheduler.client.report [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.299232] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f2bcf9-531b-abdd-f7ad-68442ebebd41, 'name': SearchDatastore_Task, 'duration_secs': 0.015468} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.300013] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc7f15bb-8dcb-4097-afd0-a220dbe32a0f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.305081] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for the task: (returnval){ [ 993.305081] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e323cd-740f-ec06-eb8b-87164bedafa3" [ 993.305081] env[63355]: _type = "Task" [ 993.305081] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.312914] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e323cd-740f-ec06-eb8b-87164bedafa3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.488996] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349968, 'name': Rename_Task, 'duration_secs': 0.267247} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.489297] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.489589] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e00e28e-cd7f-479e-95c3-fa576f2946dd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.496331] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 993.496331] env[63355]: value = "task-1349969" [ 993.496331] env[63355]: _type = "Task" [ 993.496331] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.504152] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349969, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.652421] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8659dcc0-d198-4399-b591-1515d88326ee tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "cd9de9b8-ad42-4dbb-b435-927738a55f3f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.426s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.752497] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.754909] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.637s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.756568] env[63355]: INFO nova.compute.claims [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.771452] env[63355]: INFO nova.scheduler.client.report [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Deleted allocations for instance 8e9ff9b7-636f-48df-9168-509d733278f8 [ 993.815283] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e323cd-740f-ec06-eb8b-87164bedafa3, 'name': SearchDatastore_Task, 'duration_secs': 0.021057} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.815643] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.815849] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 9ee3001b-8bf0-43ab-996a-a68dad57d8e6/9ee3001b-8bf0-43ab-996a-a68dad57d8e6.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 993.816144] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-949cfcbe-cf0d-4b1d-9451-147bdb1b601b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.824332] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for the task: (returnval){ [ 993.824332] env[63355]: value = "task-1349970" [ 993.824332] env[63355]: _type = "Task" [ 993.824332] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.834174] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349970, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.007178] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349969, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.278916] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d811099f-3a25-4b5f-b611-70c1325ccb2f tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e9ff9b7-636f-48df-9168-509d733278f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.155s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.336251] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349970, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.506910] env[63355]: DEBUG oslo_vmware.api [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1349969, 'name': PowerOnVM_Task, 'duration_secs': 0.909604} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.507232] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.507450] env[63355]: INFO nova.compute.manager [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Took 8.34 seconds to spawn the instance on the hypervisor. [ 994.507636] env[63355]: DEBUG nova.compute.manager [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 994.508488] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342110fa-f3da-4953-b23f-5b380fda1239 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.835888] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349970, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.74254} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.838282] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 9ee3001b-8bf0-43ab-996a-a68dad57d8e6/9ee3001b-8bf0-43ab-996a-a68dad57d8e6.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 994.838515] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 994.838961] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14b28ab8-eb05-422c-b99d-1a4a83362fcc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.845490] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for the task: (returnval){ [ 994.845490] env[63355]: value = "task-1349971" [ 994.845490] env[63355]: _type = "Task" [ 994.845490] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.856269] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349971, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.926250] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61cacc4-c594-4ab1-8c07-cec089384340 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.934303] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0389d70b-f421-4484-a455-f6a34ce78b25 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.966057] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34983dec-9aee-4cfc-9bcc-918746560cfb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.973985] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171133ff-7cfd-4298-9cc8-847773fe9337 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.987985] env[63355]: DEBUG nova.compute.provider_tree [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 995.027980] env[63355]: INFO nova.compute.manager [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Took 21.33 seconds to build instance. [ 995.358576] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349971, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093104} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.358960] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 995.360113] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f579702e-9128-4a34-ae62-4dc1e87ff37c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.394075] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 9ee3001b-8bf0-43ab-996a-a68dad57d8e6/9ee3001b-8bf0-43ab-996a-a68dad57d8e6.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.394518] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39632465-e666-4281-8569-d52a7b2065c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.421709] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for the task: (returnval){ [ 995.421709] env[63355]: value = "task-1349972" [ 995.421709] env[63355]: _type = "Task" [ 995.421709] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.429637] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349972, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.509197] env[63355]: ERROR nova.scheduler.client.report [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [req-3dfec783-34c3-4b33-a011-918a4d86a0c6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 47abb610-db7e-4770-911d-187dd075ef8b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3dfec783-34c3-4b33-a011-918a4d86a0c6"}]} [ 995.527844] env[63355]: DEBUG nova.scheduler.client.report [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Refreshing inventories for resource provider 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 995.529883] env[63355]: DEBUG oslo_concurrency.lockutils [None req-237b4783-b9d1-4d17-b88b-91664e169ce0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.852s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.541723] env[63355]: DEBUG nova.scheduler.client.report [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Updating ProviderTree inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 995.541955] env[63355]: DEBUG nova.compute.provider_tree [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 995.555187] env[63355]: DEBUG nova.scheduler.client.report [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Refreshing aggregate associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, aggregates: None {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 995.575055] env[63355]: DEBUG nova.scheduler.client.report [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Refreshing trait associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 995.724238] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42998f38-3e6c-4e6b-9c44-cbb6328a4c80 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.732582] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5cb7d4-ec1c-4a6a-bd72-62ebc23ff7b3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.765320] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf7801a-da3c-492b-bd6b-8267df964c39 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.773496] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae39846c-22f8-4bd1-b55b-333f71ad3a48 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.787829] env[63355]: DEBUG nova.compute.provider_tree [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 995.932472] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349972, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.320084] env[63355]: DEBUG nova.scheduler.client.report [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Updated inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b with generation 105 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 996.320390] env[63355]: DEBUG nova.compute.provider_tree [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Updating resource provider 47abb610-db7e-4770-911d-187dd075ef8b generation from 105 to 106 during operation: update_inventory {{(pid=63355) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 996.320580] env[63355]: DEBUG nova.compute.provider_tree [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 996.433305] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349972, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.826132] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.071s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.827670] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.231s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.827938] env[63355]: DEBUG nova.objects.instance [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lazy-loading 'resources' on Instance uuid ecbacdb3-516e-43c3-96e3-4961b76565ca {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.933481] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349972, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.330717] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquiring lock "802f6e51-5807-4565-8eaa-938ea51aeeaa" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.331044] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "802f6e51-5807-4565-8eaa-938ea51aeeaa" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.435021] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349972, 'name': ReconfigVM_Task, 'duration_secs': 1.558795} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.435333] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 9ee3001b-8bf0-43ab-996a-a68dad57d8e6/9ee3001b-8bf0-43ab-996a-a68dad57d8e6.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.435962] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3b92b5c-4c93-42e9-a9d9-08ae3007897d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.441752] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for the task: (returnval){ [ 997.441752] env[63355]: value = "task-1349973" [ 997.441752] env[63355]: _type = "Task" [ 997.441752] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.452312] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349973, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.479478] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49a5c27-90c1-4119-99c4-dc5ddd2e4456 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.487314] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e01a04-e687-49ab-8cb5-76272f0738f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.522257] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64506f54-4613-4c61-8bca-ce568c10aceb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.530613] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41a29e5-1ac9-4493-a69b-8930638da8ba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.544743] env[63355]: DEBUG nova.compute.provider_tree [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.836492] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "802f6e51-5807-4565-8eaa-938ea51aeeaa" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.837108] env[63355]: DEBUG nova.compute.manager [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 997.954948] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349973, 'name': Rename_Task, 'duration_secs': 0.271956} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.955260] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 997.955525] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ebb0663-40e4-422a-897f-e5c1b0d37383 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.963428] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for the task: (returnval){ [ 997.963428] env[63355]: value = "task-1349974" [ 997.963428] env[63355]: _type = "Task" [ 997.963428] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.973916] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349974, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.048323] env[63355]: DEBUG nova.scheduler.client.report [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.342344] env[63355]: DEBUG nova.compute.utils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 998.344224] env[63355]: DEBUG nova.compute.manager [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 998.344420] env[63355]: DEBUG nova.network.neutron [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 998.398297] env[63355]: DEBUG nova.policy [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7b86803bfa804a37b997c31937577e8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '212d6f93c6954d46b4b1ee70ef934938', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 998.475660] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349974, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.553633] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.726s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.556345] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.475s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.556703] env[63355]: DEBUG nova.objects.instance [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lazy-loading 'resources' on Instance uuid c630c5fe-6907-4952-9807-6e59bd1cc9e1 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.675071] env[63355]: DEBUG nova.network.neutron [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Successfully created port: 40f9f71e-902f-4420-a8b7-58e894a6939c {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 998.812811] env[63355]: DEBUG oslo_vmware.rw_handles [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524bce97-3be4-2e76-81b6-8c4ca3306e6f/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 998.813942] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3342718-3f7a-47fa-b4e0-e35a60c648f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.820612] env[63355]: DEBUG oslo_vmware.rw_handles [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524bce97-3be4-2e76-81b6-8c4ca3306e6f/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 998.820796] env[63355]: ERROR oslo_vmware.rw_handles [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524bce97-3be4-2e76-81b6-8c4ca3306e6f/disk-0.vmdk due to incomplete transfer. [ 998.821059] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a1a5b8d8-a27a-4665-91e5-16b39e2a378b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.828526] env[63355]: DEBUG oslo_vmware.rw_handles [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524bce97-3be4-2e76-81b6-8c4ca3306e6f/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 998.828735] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Uploaded image dfe52617-1814-4a03-b7f0-4f5b8cfa9c9c to the Glance image server {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 998.830878] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 998.831156] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a0cf6a1f-94a8-4d2a-911e-f0918e4074c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.836897] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 998.836897] env[63355]: value = "task-1349975" [ 998.836897] env[63355]: _type = "Task" [ 998.836897] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.845297] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349975, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.846917] env[63355]: DEBUG nova.compute.manager [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 998.905352] env[63355]: DEBUG oslo_vmware.rw_handles [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52399965-bfc6-b4c7-b8e9-12168bc4480e/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 998.906658] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e713373e-4777-4e15-a60b-5ae664bbf9bb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.913270] env[63355]: DEBUG oslo_vmware.rw_handles [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52399965-bfc6-b4c7-b8e9-12168bc4480e/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 998.913469] env[63355]: ERROR oslo_vmware.rw_handles [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52399965-bfc6-b4c7-b8e9-12168bc4480e/disk-0.vmdk due to incomplete transfer. [ 998.913746] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5f5672a3-75a1-4a2c-a538-4931d86177d4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.926442] env[63355]: DEBUG oslo_vmware.rw_handles [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52399965-bfc6-b4c7-b8e9-12168bc4480e/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 998.926442] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Uploaded image c58460ed-ef70-483a-87f9-6152e2a0d383 to the Glance image server {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 998.926442] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 998.926442] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-875629c7-bbfd-479e-9118-1837c5e5e1bb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.932454] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 998.932454] env[63355]: value = "task-1349976" [ 998.932454] env[63355]: _type = "Task" [ 998.932454] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.941706] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349976, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.973868] env[63355]: DEBUG oslo_vmware.api [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349974, 'name': PowerOnVM_Task, 'duration_secs': 0.97443} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.974156] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 998.974351] env[63355]: INFO nova.compute.manager [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Took 10.32 seconds to spawn the instance on the hypervisor. [ 998.974576] env[63355]: DEBUG nova.compute.manager [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.975367] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa20ec2-32e6-40e0-84f7-8b664fee473d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.072967] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fe5d19d1-3c65-4873-827c-87517aed6cf6 tempest-ServerActionsV293TestJSON-989167322 tempest-ServerActionsV293TestJSON-989167322-project-member] Lock "ecbacdb3-516e-43c3-96e3-4961b76565ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.688s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.213694] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1820da-cc22-4401-89b0-120a06a419f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.224448] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c58cc37-caf7-44a9-b0d3-a7d3b2507b29 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.252800] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec35689e-8222-4df0-8771-f998e50496c0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.262497] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff79447-3d8a-451a-91fa-d3eebf1cf0c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.284320] env[63355]: DEBUG nova.compute.provider_tree [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.362480] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349975, 'name': Destroy_Task} progress is 33%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.443522] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349976, 'name': Destroy_Task} progress is 33%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.494846] env[63355]: INFO nova.compute.manager [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Took 24.21 seconds to build instance. [ 999.788871] env[63355]: DEBUG nova.scheduler.client.report [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 999.850260] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349975, 'name': Destroy_Task, 'duration_secs': 0.819957} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.853514] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Destroyed the VM [ 999.853892] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 999.854395] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7902b4b2-f207-48bf-b8a9-041a3e4c2279 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.857724] env[63355]: DEBUG nova.compute.manager [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 999.873565] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.873822] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.875366] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.875711] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.877186] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 999.877186] env[63355]: value = "task-1349977" [ 999.877186] env[63355]: _type = "Task" [ 999.877186] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.887907] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.888265] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.888466] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.888674] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.888827] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.888980] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.889500] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.889717] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.889898] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.890103] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.890322] env[63355]: DEBUG nova.virt.hardware [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.892233] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4256da75-5084-49ab-8207-2adca3a16d47 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.898699] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349977, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.904388] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba93526-8620-443c-8b9e-b7169f3a40b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.945612] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349976, 'name': Destroy_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.997402] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5fbe13df-f886-4127-8acd-8b54dc75b369 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.722s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.293946] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.738s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.296994] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.825s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.296994] env[63355]: DEBUG nova.objects.instance [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lazy-loading 'resources' on Instance uuid f5e62ce1-40b7-4648-a4a6-068ff06eaf9b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.310090] env[63355]: INFO nova.scheduler.client.report [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Deleted allocations for instance c630c5fe-6907-4952-9807-6e59bd1cc9e1 [ 1000.354324] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.355224] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.355224] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.355224] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.355511] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.361019] env[63355]: INFO nova.compute.manager [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Terminating instance [ 1000.361019] env[63355]: DEBUG nova.compute.manager [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.361019] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.361688] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57b5b8e-ed97-4506-978a-99673479c9df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.371565] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.377242] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec89b18a-c6ce-42c1-abc3-e3bd625fdf79 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.379849] env[63355]: DEBUG nova.compute.manager [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.383154] env[63355]: DEBUG nova.compute.manager [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.392586] env[63355]: DEBUG oslo_vmware.api [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 1000.392586] env[63355]: value = "task-1349978" [ 1000.392586] env[63355]: _type = "Task" [ 1000.392586] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.399815] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349977, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.407817] env[63355]: DEBUG oslo_vmware.api [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.446041] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349976, 'name': Destroy_Task, 'duration_secs': 1.017989} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.446701] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Destroyed the VM [ 1000.450042] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1000.450042] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-542e4cf5-4cc0-4fcd-8126-15e1636c7fe8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.454756] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1000.454756] env[63355]: value = "task-1349979" [ 1000.454756] env[63355]: _type = "Task" [ 1000.454756] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.464149] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349979, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.796289] env[63355]: DEBUG nova.network.neutron [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Successfully updated port: 40f9f71e-902f-4420-a8b7-58e894a6939c {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1000.820742] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1a8c1578-9309-43b6-833a-17c284a70578 tempest-ServerAddressesNegativeTestJSON-1065584175 tempest-ServerAddressesNegativeTestJSON-1065584175-project-member] Lock "c630c5fe-6907-4952-9807-6e59bd1cc9e1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.668s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.905306] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.923235] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.923498] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.924988] env[63355]: DEBUG oslo_vmware.api [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349977, 'name': RemoveSnapshot_Task, 'duration_secs': 0.548829} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.925936] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.932194] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1000.932194] env[63355]: INFO nova.compute.manager [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Took 16.09 seconds to snapshot the instance on the hypervisor. [ 1000.932930] env[63355]: DEBUG oslo_vmware.api [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349978, 'name': PowerOffVM_Task, 'duration_secs': 0.257557} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.933205] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.933365] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.933649] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d9dd41e-5435-490d-ab03-fd7883996520 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.972615] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349979, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.002476] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.002709] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.002897] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Deleting the datastore file [datastore2] 8e3ec9d3-bc22-4e39-ad7c-93268dd59020 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.003285] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e754c39-7022-4ca0-9384-a9bd24a34828 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.013117] env[63355]: DEBUG oslo_vmware.api [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for the task: (returnval){ [ 1001.013117] env[63355]: value = "task-1349981" [ 1001.013117] env[63355]: _type = "Task" [ 1001.013117] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.024278] env[63355]: DEBUG oslo_vmware.api [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349981, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.099512] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a07bc0-8659-4b1e-9c79-ef91d910b660 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.112760] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0259e7-6888-4a4d-a75b-6e5fb2ec803e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.148682] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ef2695-de19-4d67-99c1-6946a54b7e50 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.156959] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9383e5-6d36-43d4-8199-1d0bd2bf8ee3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.171294] env[63355]: DEBUG nova.compute.provider_tree [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.302181] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquiring lock "refresh_cache-9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.302340] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquired lock "refresh_cache-9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.302566] env[63355]: DEBUG nova.network.neutron [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.351728] env[63355]: DEBUG nova.compute.manager [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Stashing vm_state: active {{(pid=63355) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1001.427526] env[63355]: DEBUG nova.compute.manager [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1001.474414] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349979, 'name': RemoveSnapshot_Task, 'duration_secs': 0.652919} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.474930] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1001.475054] env[63355]: DEBUG nova.compute.manager [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.476224] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c6b4bc-c8cd-4d8a-a0a1-de6b0a6e25bc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.499588] env[63355]: DEBUG nova.compute.manager [None req-df1e2ba6-a805-4d2e-8c80-e65397763aa0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Found 2 images (rotation: 2) {{(pid=63355) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1001.526780] env[63355]: DEBUG oslo_vmware.api [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Task: {'id': task-1349981, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.424917} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.526780] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.526780] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.526780] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.529243] env[63355]: INFO nova.compute.manager [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1001.532021] env[63355]: DEBUG oslo.service.loopingcall [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.532021] env[63355]: DEBUG nova.compute.manager [-] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1001.532021] env[63355]: DEBUG nova.network.neutron [-] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1001.680113] env[63355]: DEBUG nova.scheduler.client.report [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.837859] env[63355]: DEBUG nova.network.neutron [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1001.877108] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.917497] env[63355]: DEBUG nova.compute.manager [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Received event network-vif-plugged-40f9f71e-902f-4420-a8b7-58e894a6939c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.917804] env[63355]: DEBUG oslo_concurrency.lockutils [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] Acquiring lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.917968] env[63355]: DEBUG oslo_concurrency.lockutils [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] Lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.918254] env[63355]: DEBUG oslo_concurrency.lockutils [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] Lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.918491] env[63355]: DEBUG nova.compute.manager [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] No waiting events found dispatching network-vif-plugged-40f9f71e-902f-4420-a8b7-58e894a6939c {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1001.918715] env[63355]: WARNING nova.compute.manager [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Received unexpected event network-vif-plugged-40f9f71e-902f-4420-a8b7-58e894a6939c for instance with vm_state building and task_state spawning. [ 1001.918986] env[63355]: DEBUG nova.compute.manager [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Received event network-changed-40f9f71e-902f-4420-a8b7-58e894a6939c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.919207] env[63355]: DEBUG nova.compute.manager [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Refreshing instance network info cache due to event network-changed-40f9f71e-902f-4420-a8b7-58e894a6939c. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1001.920113] env[63355]: DEBUG oslo_concurrency.lockutils [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] Acquiring lock "refresh_cache-9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.952197] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.990451] env[63355]: DEBUG nova.network.neutron [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Updating instance_info_cache with network_info: [{"id": "40f9f71e-902f-4420-a8b7-58e894a6939c", "address": "fa:16:3e:49:7a:ec", "network": {"id": "e5c4e7ac-d7bc-4bed-ab6d-665823e33ff2", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-469934042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212d6f93c6954d46b4b1ee70ef934938", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40f9f71e-90", "ovs_interfaceid": "40f9f71e-902f-4420-a8b7-58e894a6939c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.999489] env[63355]: INFO nova.compute.manager [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Shelve offloading [ 1002.001333] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1002.001790] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5d65415-0310-4dad-a74e-df127ae04307 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.009984] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1002.009984] env[63355]: value = "task-1349982" [ 1002.009984] env[63355]: _type = "Task" [ 1002.009984] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.020700] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] VM already powered off {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1002.020823] env[63355]: DEBUG nova.compute.manager [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.022144] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1ef8ed-d759-42cd-8522-1f74537127bf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.029440] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.029661] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.029844] env[63355]: DEBUG nova.network.neutron [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1002.142361] env[63355]: DEBUG nova.compute.manager [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.146707] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9e6d84-08f3-4ed4-9443-322eebde1cd1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.153931] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "fa748748-31ec-431a-a628-5ea179e26fc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.154472] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "fa748748-31ec-431a-a628-5ea179e26fc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.185059] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.889s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.188857] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.283s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.190226] env[63355]: INFO nova.compute.claims [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.217724] env[63355]: INFO nova.scheduler.client.report [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleted allocations for instance f5e62ce1-40b7-4648-a4a6-068ff06eaf9b [ 1002.495028] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Releasing lock "refresh_cache-9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.495489] env[63355]: DEBUG nova.compute.manager [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Instance network_info: |[{"id": "40f9f71e-902f-4420-a8b7-58e894a6939c", "address": "fa:16:3e:49:7a:ec", "network": {"id": "e5c4e7ac-d7bc-4bed-ab6d-665823e33ff2", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-469934042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212d6f93c6954d46b4b1ee70ef934938", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40f9f71e-90", "ovs_interfaceid": "40f9f71e-902f-4420-a8b7-58e894a6939c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1002.495907] env[63355]: DEBUG oslo_concurrency.lockutils [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] Acquired lock "refresh_cache-9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.496184] env[63355]: DEBUG nova.network.neutron [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Refreshing network info cache for port 40f9f71e-902f-4420-a8b7-58e894a6939c {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1002.497827] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:7a:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '08e9585e-6186-4788-9fd9-24174ce45a6f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40f9f71e-902f-4420-a8b7-58e894a6939c', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1002.509650] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Creating folder: Project (212d6f93c6954d46b4b1ee70ef934938). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1002.510993] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d303aa4-6eb6-44e3-8f50-a19123eda6c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.522642] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Created folder: Project (212d6f93c6954d46b4b1ee70ef934938) in parent group-v287607. [ 1002.522845] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Creating folder: Instances. Parent ref: group-v287732. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1002.523130] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8a68c60b-aaa3-40a9-b2e4-fffa18678678 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.532447] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Created folder: Instances in parent group-v287732. [ 1002.532687] env[63355]: DEBUG oslo.service.loopingcall [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.532879] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1002.535028] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08da69af-67e3-42e3-aa39-60ac36e67e46 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.554494] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1002.554494] env[63355]: value = "task-1349985" [ 1002.554494] env[63355]: _type = "Task" [ 1002.554494] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.557899] env[63355]: DEBUG nova.network.neutron [-] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.562449] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349985, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.657036] env[63355]: DEBUG nova.compute.manager [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1002.667373] env[63355]: INFO nova.compute.manager [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] instance snapshotting [ 1002.667980] env[63355]: DEBUG nova.objects.instance [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'flavor' on Instance uuid 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.725695] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f1aea3d9-61e6-4597-841a-381fa80a672c tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "f5e62ce1-40b7-4648-a4a6-068ff06eaf9b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.222s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.796933] env[63355]: DEBUG nova.network.neutron [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updating instance_info_cache with network_info: [{"id": "43ebe841-d4c8-446a-981c-519bef977228", "address": "fa:16:3e:97:62:7e", "network": {"id": "e19b4619-6e81-4205-838d-9ee7e7ae44d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1968447346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "744ce7dc0ab4463fa461cbac87e51904", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43ebe841-d4", "ovs_interfaceid": "43ebe841-d4c8-446a-981c-519bef977228", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.829992] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquiring lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.830345] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.830522] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquiring lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.830711] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.830888] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.834558] env[63355]: INFO nova.compute.manager [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Terminating instance [ 1002.836539] env[63355]: DEBUG nova.compute.manager [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1002.836736] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1002.837699] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16663815-f550-4038-803a-8dfb1e27e955 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.846118] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1002.846452] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b058444-6525-41ba-be1e-43b6f87cb9bc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.852408] env[63355]: DEBUG oslo_vmware.api [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for the task: (returnval){ [ 1002.852408] env[63355]: value = "task-1349986" [ 1002.852408] env[63355]: _type = "Task" [ 1002.852408] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.860580] env[63355]: DEBUG oslo_vmware.api [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349986, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.064469] env[63355]: INFO nova.compute.manager [-] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Took 1.53 seconds to deallocate network for instance. [ 1003.064716] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1349985, 'name': CreateVM_Task, 'duration_secs': 0.374173} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.067620] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1003.067620] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.067620] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.067620] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1003.067836] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14c77734-73d6-49a2-a436-0beee15f1e85 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.075854] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for the task: (returnval){ [ 1003.075854] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f5322d-81e5-48bf-2422-95376d6f8109" [ 1003.075854] env[63355]: _type = "Task" [ 1003.075854] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.085502] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f5322d-81e5-48bf-2422-95376d6f8109, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.174389] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1e9b2a-303a-4d51-b7b2-9b7dc8c62b2e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.196100] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.197085] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634c7fb4-9861-4d7d-a994-23bf8d48e46d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.302569] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Releasing lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.364114] env[63355]: DEBUG oslo_vmware.api [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349986, 'name': PowerOffVM_Task, 'duration_secs': 0.201625} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.364114] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.364114] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.364114] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-208db885-6207-4de9-a111-acba1b3f2a43 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.420476] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81445ed6-c190-403d-89e8-69e347d22061 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.425533] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1003.425748] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1003.425925] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Deleting the datastore file [datastore1] 9ee3001b-8bf0-43ab-996a-a68dad57d8e6 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.427786] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7af52eaa-a6c8-4804-8850-16d224ed4951 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.430613] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9f1702-6ee1-4c08-9ff3-ed835895275c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.438193] env[63355]: DEBUG oslo_vmware.api [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for the task: (returnval){ [ 1003.438193] env[63355]: value = "task-1349988" [ 1003.438193] env[63355]: _type = "Task" [ 1003.438193] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.464406] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55469d31-5a25-4b86-acd1-fba8f86d5a64 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.471855] env[63355]: DEBUG oslo_vmware.api [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349988, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.481308] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1074f9-fe88-464e-a548-a71522081020 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.497188] env[63355]: DEBUG nova.compute.provider_tree [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1003.549833] env[63355]: DEBUG nova.network.neutron [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Updated VIF entry in instance network info cache for port 40f9f71e-902f-4420-a8b7-58e894a6939c. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1003.551426] env[63355]: DEBUG nova.network.neutron [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Updating instance_info_cache with network_info: [{"id": "40f9f71e-902f-4420-a8b7-58e894a6939c", "address": "fa:16:3e:49:7a:ec", "network": {"id": "e5c4e7ac-d7bc-4bed-ab6d-665823e33ff2", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-469934042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212d6f93c6954d46b4b1ee70ef934938", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40f9f71e-90", "ovs_interfaceid": "40f9f71e-902f-4420-a8b7-58e894a6939c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.573785] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.587285] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f5322d-81e5-48bf-2422-95376d6f8109, 'name': SearchDatastore_Task, 'duration_secs': 0.008596} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.587600] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.588078] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.588209] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.588370] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.588595] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.588861] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a6cc5905-af7d-4fbf-b8d4-15a79926e98a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.598882] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.599088] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.600480] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd1995b6-29ee-4417-9731-95c7f6d18cd2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.605919] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for the task: (returnval){ [ 1003.605919] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52611899-6de4-d579-1e24-54c8bcdea6f7" [ 1003.605919] env[63355]: _type = "Task" [ 1003.605919] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.616349] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52611899-6de4-d579-1e24-54c8bcdea6f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.620525] env[63355]: DEBUG nova.compute.manager [req-95e6c32a-1d2b-4b15-b6b9-83b707dbc68b req-ed93d253-31ae-4ee0-84e0-ddeb5dbb4675 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received event network-vif-unplugged-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.620928] env[63355]: DEBUG oslo_concurrency.lockutils [req-95e6c32a-1d2b-4b15-b6b9-83b707dbc68b req-ed93d253-31ae-4ee0-84e0-ddeb5dbb4675 service nova] Acquiring lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.621296] env[63355]: DEBUG oslo_concurrency.lockutils [req-95e6c32a-1d2b-4b15-b6b9-83b707dbc68b req-ed93d253-31ae-4ee0-84e0-ddeb5dbb4675 service nova] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.621568] env[63355]: DEBUG oslo_concurrency.lockutils [req-95e6c32a-1d2b-4b15-b6b9-83b707dbc68b req-ed93d253-31ae-4ee0-84e0-ddeb5dbb4675 service nova] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.621818] env[63355]: DEBUG nova.compute.manager [req-95e6c32a-1d2b-4b15-b6b9-83b707dbc68b req-ed93d253-31ae-4ee0-84e0-ddeb5dbb4675 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] No waiting events found dispatching network-vif-unplugged-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1003.622173] env[63355]: WARNING nova.compute.manager [req-95e6c32a-1d2b-4b15-b6b9-83b707dbc68b req-ed93d253-31ae-4ee0-84e0-ddeb5dbb4675 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received unexpected event network-vif-unplugged-43ebe841-d4c8-446a-981c-519bef977228 for instance with vm_state shelved and task_state shelving_offloading. [ 1003.675630] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.676628] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cc797f-f85d-41b6-8cbe-5df4dd1fd4e3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.683788] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.684054] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3888baf0-fd7a-413d-b113-c6c0da052ebb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.712577] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1003.712853] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7ad5ef80-5e5e-43fc-a5ea-cd6a15bf3a74 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.719479] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1003.719479] env[63355]: value = "task-1349990" [ 1003.719479] env[63355]: _type = "Task" [ 1003.719479] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.727200] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349990, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.786547] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1003.786806] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1003.787034] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Deleting the datastore file [datastore1] 7c153109-b814-4e11-b4f9-7b8cebb853d4 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.787331] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15041462-a083-4eec-aeb4-6d64cf19026c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.793845] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1003.793845] env[63355]: value = "task-1349991" [ 1003.793845] env[63355]: _type = "Task" [ 1003.793845] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.801351] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349991, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.971448] env[63355]: DEBUG oslo_vmware.api [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Task: {'id': task-1349988, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127588} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.971797] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.971991] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1003.972189] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1003.972373] env[63355]: INFO nova.compute.manager [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1003.972624] env[63355]: DEBUG oslo.service.loopingcall [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.972831] env[63355]: DEBUG nova.compute.manager [-] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1003.972917] env[63355]: DEBUG nova.network.neutron [-] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1004.005960] env[63355]: DEBUG nova.compute.manager [req-af98a3c5-1d72-44f3-ad08-47273a7b2fcf req-814efe94-2a99-4ce1-a3d3-9d22dc4b8704 service nova] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Received event network-vif-deleted-b134b9dc-a12b-48b4-8051-39ec142ce649 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.021719] env[63355]: ERROR nova.scheduler.client.report [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [req-d46bdbb7-6b5a-41c4-a6e0-a3628917d362] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 47abb610-db7e-4770-911d-187dd075ef8b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d46bdbb7-6b5a-41c4-a6e0-a3628917d362"}]} [ 1004.037766] env[63355]: DEBUG nova.scheduler.client.report [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Refreshing inventories for resource provider 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1004.054424] env[63355]: DEBUG oslo_concurrency.lockutils [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] Releasing lock "refresh_cache-9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.054730] env[63355]: DEBUG nova.compute.manager [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received event network-changed-aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.054916] env[63355]: DEBUG nova.compute.manager [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing instance network info cache due to event network-changed-aa5a40b6-5d12-4d62-88d2-32a26a65a241. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1004.055155] env[63355]: DEBUG oslo_concurrency.lockutils [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] Acquiring lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.055332] env[63355]: DEBUG oslo_concurrency.lockutils [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] Acquired lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.055506] env[63355]: DEBUG nova.network.neutron [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing network info cache for port aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1004.058963] env[63355]: DEBUG nova.scheduler.client.report [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Updating ProviderTree inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1004.059259] env[63355]: DEBUG nova.compute.provider_tree [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.074261] env[63355]: DEBUG nova.scheduler.client.report [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Refreshing aggregate associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, aggregates: None {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1004.092684] env[63355]: DEBUG nova.scheduler.client.report [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Refreshing trait associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1004.118494] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52611899-6de4-d579-1e24-54c8bcdea6f7, 'name': SearchDatastore_Task, 'duration_secs': 0.022277} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.119367] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96fd8962-4903-487e-bbd4-6d481e5b338e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.126869] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for the task: (returnval){ [ 1004.126869] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524beb57-5cf0-09b5-5ff7-6b8ab88b9d12" [ 1004.126869] env[63355]: _type = "Task" [ 1004.126869] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.134810] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524beb57-5cf0-09b5-5ff7-6b8ab88b9d12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.234222] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349990, 'name': CreateSnapshot_Task, 'duration_secs': 0.474086} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.234587] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1004.236576] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e7b899-8da2-4280-b6d0-d5c7f04b7259 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.277988] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86f71c2-081a-4921-8bbe-6ce258b4ef79 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.287053] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4409a17-6392-4741-84f6-38af40edbc8f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.321428] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8024f7ee-154f-46ae-aa21-6709f9b615ca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.332064] env[63355]: DEBUG oslo_vmware.api [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1349991, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206742} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.333271] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c023df19-593a-4430-80eb-170351aad534 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.336963] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.337191] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.337378] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.349508] env[63355]: DEBUG nova.compute.provider_tree [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.356319] env[63355]: INFO nova.scheduler.client.report [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Deleted allocations for instance 7c153109-b814-4e11-b4f9-7b8cebb853d4 [ 1004.639950] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524beb57-5cf0-09b5-5ff7-6b8ab88b9d12, 'name': SearchDatastore_Task, 'duration_secs': 0.010661} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.639950] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.640589] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7/9bf49da7-df44-4f26-ac7d-d3a4dab24ce7.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.640756] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70960b4d-afe6-4896-b6fc-4b587a549611 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.649790] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for the task: (returnval){ [ 1004.649790] env[63355]: value = "task-1349992" [ 1004.649790] env[63355]: _type = "Task" [ 1004.649790] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.660134] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349992, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.742428] env[63355]: DEBUG nova.network.neutron [-] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.758364] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1004.759236] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-20095f0e-90e1-47d7-8229-cbf7790e1763 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.769013] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1004.769013] env[63355]: value = "task-1349993" [ 1004.769013] env[63355]: _type = "Task" [ 1004.769013] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.780964] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349993, 'name': CloneVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.783387] env[63355]: DEBUG nova.network.neutron [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updated VIF entry in instance network info cache for port aa5a40b6-5d12-4d62-88d2-32a26a65a241. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1004.783387] env[63355]: DEBUG nova.network.neutron [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [{"id": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "address": "fa:16:3e:08:9e:2a", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5a40b6-5d", "ovs_interfaceid": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.860841] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.886706] env[63355]: DEBUG nova.scheduler.client.report [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Updated inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b with generation 112 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1004.887533] env[63355]: DEBUG nova.compute.provider_tree [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Updating resource provider 47abb610-db7e-4770-911d-187dd075ef8b generation from 112 to 113 during operation: update_inventory {{(pid=63355) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1004.887533] env[63355]: DEBUG nova.compute.provider_tree [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.163181] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349992, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.247629] env[63355]: INFO nova.compute.manager [-] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Took 1.27 seconds to deallocate network for instance. [ 1005.279439] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349993, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.286490] env[63355]: DEBUG oslo_concurrency.lockutils [req-d5b1937b-dfd6-4b73-a143-b1a421a6b944 req-932abfe7-06ab-4a38-8a21-e122c490ea63 service nova] Releasing lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.392887] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.204s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.393468] env[63355]: DEBUG nova.compute.manager [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1005.396181] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.470s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.397707] env[63355]: INFO nova.compute.claims [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.661541] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349992, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556592} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.661843] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7/9bf49da7-df44-4f26-ac7d-d3a4dab24ce7.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1005.662079] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.662374] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5585f468-a511-420b-a359-9feb836ca10f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.670632] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for the task: (returnval){ [ 1005.670632] env[63355]: value = "task-1349994" [ 1005.670632] env[63355]: _type = "Task" [ 1005.670632] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.680870] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349994, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.728504] env[63355]: DEBUG nova.compute.manager [req-3e014ead-ad80-4559-9ff7-4263580f69df req-c8555d3a-74a0-46c4-88b7-5b28a9840cc7 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received event network-changed-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.728734] env[63355]: DEBUG nova.compute.manager [req-3e014ead-ad80-4559-9ff7-4263580f69df req-c8555d3a-74a0-46c4-88b7-5b28a9840cc7 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Refreshing instance network info cache due to event network-changed-43ebe841-d4c8-446a-981c-519bef977228. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1005.729109] env[63355]: DEBUG oslo_concurrency.lockutils [req-3e014ead-ad80-4559-9ff7-4263580f69df req-c8555d3a-74a0-46c4-88b7-5b28a9840cc7 service nova] Acquiring lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.729109] env[63355]: DEBUG oslo_concurrency.lockutils [req-3e014ead-ad80-4559-9ff7-4263580f69df req-c8555d3a-74a0-46c4-88b7-5b28a9840cc7 service nova] Acquired lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.729413] env[63355]: DEBUG nova.network.neutron [req-3e014ead-ad80-4559-9ff7-4263580f69df req-c8555d3a-74a0-46c4-88b7-5b28a9840cc7 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Refreshing network info cache for port 43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1005.756020] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.781038] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349993, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.902422] env[63355]: DEBUG nova.compute.utils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.903874] env[63355]: DEBUG nova.compute.manager [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1005.904089] env[63355]: DEBUG nova.network.neutron [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1005.946381] env[63355]: DEBUG nova.policy [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3721c4a73df54ae9b44110cb9e8590a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc547065748241e8ac7b6c499ddaea66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1006.181627] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349994, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.265861} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.181981] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1006.182940] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c69d100-d4c2-4df3-ab62-f29f13a724be {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.206805] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7/9bf49da7-df44-4f26-ac7d-d3a4dab24ce7.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.207133] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bae47ae1-4a90-4f86-a1c4-90e80775ca72 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.237520] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for the task: (returnval){ [ 1006.237520] env[63355]: value = "task-1349995" [ 1006.237520] env[63355]: _type = "Task" [ 1006.237520] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.251471] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349995, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.289676] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349993, 'name': CloneVM_Task} progress is 95%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.303554] env[63355]: DEBUG nova.network.neutron [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Successfully created port: 29b48485-2c25-460b-a265-d26ccf3a337c {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1006.409484] env[63355]: DEBUG nova.compute.manager [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1006.571347] env[63355]: DEBUG nova.network.neutron [req-3e014ead-ad80-4559-9ff7-4263580f69df req-c8555d3a-74a0-46c4-88b7-5b28a9840cc7 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updated VIF entry in instance network info cache for port 43ebe841-d4c8-446a-981c-519bef977228. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1006.571862] env[63355]: DEBUG nova.network.neutron [req-3e014ead-ad80-4559-9ff7-4263580f69df req-c8555d3a-74a0-46c4-88b7-5b28a9840cc7 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updating instance_info_cache with network_info: [{"id": "43ebe841-d4c8-446a-981c-519bef977228", "address": "fa:16:3e:97:62:7e", "network": {"id": "e19b4619-6e81-4205-838d-9ee7e7ae44d8", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1968447346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "744ce7dc0ab4463fa461cbac87e51904", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap43ebe841-d4", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.632147] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0967dd8-d82a-46f2-b630-3c587399c7e5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.641019] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7946bef7-c8f4-40d7-a5e8-2bf848da4f53 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.673240] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc49afe6-6414-4a9c-8723-f234c7f08709 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.679346] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342add00-6ea2-4dc7-96e3-12f6a58e56b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.694499] env[63355]: DEBUG nova.compute.provider_tree [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.747848] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349995, 'name': ReconfigVM_Task, 'duration_secs': 0.409452} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.748188] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7/9bf49da7-df44-4f26-ac7d-d3a4dab24ce7.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.748820] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc2f4d3a-6ddf-48e7-8055-dbeb45f532b4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.756662] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for the task: (returnval){ [ 1006.756662] env[63355]: value = "task-1349996" [ 1006.756662] env[63355]: _type = "Task" [ 1006.756662] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.764746] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349996, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.783648] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1349993, 'name': CloneVM_Task, 'duration_secs': 1.872652} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.783941] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Created linked-clone VM from snapshot [ 1006.785023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8519f27-80e0-498b-8066-dc5d38872761 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.793846] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Uploading image 3061ee1f-5dc1-4bba-9208-313cf0785cd6 {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1006.820192] env[63355]: DEBUG oslo_vmware.rw_handles [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1006.820192] env[63355]: value = "vm-287736" [ 1006.820192] env[63355]: _type = "VirtualMachine" [ 1006.820192] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1006.820496] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-07cccdc7-e6ff-4f96-a570-b410cb9c9bff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.830399] env[63355]: DEBUG oslo_vmware.rw_handles [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lease: (returnval){ [ 1006.830399] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528bd827-0eca-08d0-e6f0-b658a7839910" [ 1006.830399] env[63355]: _type = "HttpNfcLease" [ 1006.830399] env[63355]: } obtained for exporting VM: (result){ [ 1006.830399] env[63355]: value = "vm-287736" [ 1006.830399] env[63355]: _type = "VirtualMachine" [ 1006.830399] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1006.830841] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the lease: (returnval){ [ 1006.830841] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528bd827-0eca-08d0-e6f0-b658a7839910" [ 1006.830841] env[63355]: _type = "HttpNfcLease" [ 1006.830841] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1006.838195] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1006.838195] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528bd827-0eca-08d0-e6f0-b658a7839910" [ 1006.838195] env[63355]: _type = "HttpNfcLease" [ 1006.838195] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1007.076101] env[63355]: DEBUG oslo_concurrency.lockutils [req-3e014ead-ad80-4559-9ff7-4263580f69df req-c8555d3a-74a0-46c4-88b7-5b28a9840cc7 service nova] Releasing lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.076514] env[63355]: DEBUG nova.compute.manager [req-3e014ead-ad80-4559-9ff7-4263580f69df req-c8555d3a-74a0-46c4-88b7-5b28a9840cc7 service nova] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Received event network-vif-deleted-f7b42e7c-9480-4b58-a4de-69fb8b6c76ce {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.201724] env[63355]: DEBUG nova.scheduler.client.report [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.205064] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.267161] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349996, 'name': Rename_Task, 'duration_secs': 0.153391} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.267532] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1007.267821] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-adc3e65f-6cdf-4df8-be5d-b512682ad5fe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.275454] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for the task: (returnval){ [ 1007.275454] env[63355]: value = "task-1349998" [ 1007.275454] env[63355]: _type = "Task" [ 1007.275454] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.283729] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349998, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.339202] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1007.339202] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528bd827-0eca-08d0-e6f0-b658a7839910" [ 1007.339202] env[63355]: _type = "HttpNfcLease" [ 1007.339202] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1007.339602] env[63355]: DEBUG oslo_vmware.rw_handles [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1007.339602] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528bd827-0eca-08d0-e6f0-b658a7839910" [ 1007.339602] env[63355]: _type = "HttpNfcLease" [ 1007.339602] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1007.340341] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61eb1343-21ce-4a83-99f2-fe6c5c1ce28b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.348594] env[63355]: DEBUG oslo_vmware.rw_handles [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5a3e6-f9e5-7f87-64cf-90f880ce5276/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1007.348781] env[63355]: DEBUG oslo_vmware.rw_handles [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5a3e6-f9e5-7f87-64cf-90f880ce5276/disk-0.vmdk for reading. {{(pid=63355) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1007.421278] env[63355]: DEBUG nova.compute.manager [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1007.445970] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.446249] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.446493] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.446693] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.446861] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.447031] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.447249] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.447490] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.447674] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.447843] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.448033] env[63355]: DEBUG nova.virt.hardware [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.448929] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f184c79-44eb-4850-b70d-d756c4cb385c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.455167] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-838035c5-04ac-47e9-ad20-61b35ce976a3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.461218] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffea4afd-38fc-4c10-9777-7884042dc749 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.710033] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.710033] env[63355]: DEBUG nova.compute.manager [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1007.710894] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.834s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.765645] env[63355]: DEBUG nova.compute.manager [req-408ea5ff-f31c-4398-89ce-15ac946dc785 req-7978cb5b-4c9e-4bde-895e-4709b8a9371a service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Received event network-vif-plugged-29b48485-2c25-460b-a265-d26ccf3a337c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.765645] env[63355]: DEBUG oslo_concurrency.lockutils [req-408ea5ff-f31c-4398-89ce-15ac946dc785 req-7978cb5b-4c9e-4bde-895e-4709b8a9371a service nova] Acquiring lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.765645] env[63355]: DEBUG oslo_concurrency.lockutils [req-408ea5ff-f31c-4398-89ce-15ac946dc785 req-7978cb5b-4c9e-4bde-895e-4709b8a9371a service nova] Lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.765645] env[63355]: DEBUG oslo_concurrency.lockutils [req-408ea5ff-f31c-4398-89ce-15ac946dc785 req-7978cb5b-4c9e-4bde-895e-4709b8a9371a service nova] Lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.765645] env[63355]: DEBUG nova.compute.manager [req-408ea5ff-f31c-4398-89ce-15ac946dc785 req-7978cb5b-4c9e-4bde-895e-4709b8a9371a service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] No waiting events found dispatching network-vif-plugged-29b48485-2c25-460b-a265-d26ccf3a337c {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1007.765645] env[63355]: WARNING nova.compute.manager [req-408ea5ff-f31c-4398-89ce-15ac946dc785 req-7978cb5b-4c9e-4bde-895e-4709b8a9371a service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Received unexpected event network-vif-plugged-29b48485-2c25-460b-a265-d26ccf3a337c for instance with vm_state building and task_state spawning. [ 1007.786833] env[63355]: DEBUG oslo_vmware.api [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349998, 'name': PowerOnVM_Task, 'duration_secs': 0.482391} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.787118] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1007.787737] env[63355]: INFO nova.compute.manager [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Took 7.93 seconds to spawn the instance on the hypervisor. [ 1007.787737] env[63355]: DEBUG nova.compute.manager [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.788920] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a493d587-13e7-4691-aa38-7b2316db95f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.859309] env[63355]: DEBUG nova.network.neutron [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Successfully updated port: 29b48485-2c25-460b-a265-d26ccf3a337c {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1008.215448] env[63355]: DEBUG nova.compute.utils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1008.225961] env[63355]: INFO nova.compute.claims [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.232636] env[63355]: DEBUG nova.compute.manager [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1008.233175] env[63355]: DEBUG nova.network.neutron [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1008.302078] env[63355]: DEBUG nova.policy [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7123da3ee74b497395f1681e460aaaca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b6963b8e16b4986a4545914b75a38ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1008.313834] env[63355]: INFO nova.compute.manager [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Took 29.22 seconds to build instance. [ 1008.362687] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "refresh_cache-cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.362967] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "refresh_cache-cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.363376] env[63355]: DEBUG nova.network.neutron [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.591276] env[63355]: DEBUG nova.network.neutron [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Successfully created port: a0dfd814-02e8-40d1-acbc-d9447b89a9d6 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1008.649512] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquiring lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.734854] env[63355]: INFO nova.compute.resource_tracker [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating resource usage from migration cab18143-dc8b-4891-851e-80d74061526f [ 1008.737883] env[63355]: DEBUG nova.compute.manager [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1008.818129] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54cd5b09-5d5a-4a92-a52b-b4cb76abdb6f tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.733s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.818505] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.169s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.818902] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquiring lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.819352] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.819606] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.831708] env[63355]: INFO nova.compute.manager [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Terminating instance [ 1008.832805] env[63355]: DEBUG nova.compute.manager [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1008.833055] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.833954] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1558670c-56fc-4c61-ba7d-725082532bc9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.843375] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1008.848205] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d5a667ba-b39f-468b-a4f2-0f60e05791ba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.854890] env[63355]: DEBUG oslo_vmware.api [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for the task: (returnval){ [ 1008.854890] env[63355]: value = "task-1349999" [ 1008.854890] env[63355]: _type = "Task" [ 1008.854890] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.867673] env[63355]: DEBUG oslo_vmware.api [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349999, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.923600] env[63355]: DEBUG nova.network.neutron [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.989272] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b559cf04-a22f-4fba-af15-9833f33ad710 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.998706] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9f1836-2f40-4682-8431-2940b2425caf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.036467] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec40315-76b6-4c7f-a12a-67e6ef4e985d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.046763] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba022b51-8b61-42d7-84f2-14be309b1d72 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.064399] env[63355]: DEBUG nova.compute.provider_tree [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.177328] env[63355]: DEBUG nova.network.neutron [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Updating instance_info_cache with network_info: [{"id": "29b48485-2c25-460b-a265-d26ccf3a337c", "address": "fa:16:3e:26:c6:46", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29b48485-2c", "ovs_interfaceid": "29b48485-2c25-460b-a265-d26ccf3a337c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.368363] env[63355]: DEBUG oslo_vmware.api [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1349999, 'name': PowerOffVM_Task, 'duration_secs': 0.271239} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.368999] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.368999] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.369248] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79493c49-4012-4e19-8ae6-c0172acae04b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.438738] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1009.438939] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1009.439061] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Deleting the datastore file [datastore2] 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.439359] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9baa0a2a-444f-457a-ab22-b09833aed43d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.450689] env[63355]: DEBUG oslo_vmware.api [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for the task: (returnval){ [ 1009.450689] env[63355]: value = "task-1350001" [ 1009.450689] env[63355]: _type = "Task" [ 1009.450689] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.460333] env[63355]: DEBUG oslo_vmware.api [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1350001, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.567263] env[63355]: DEBUG nova.scheduler.client.report [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.681268] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "refresh_cache-cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.682041] env[63355]: DEBUG nova.compute.manager [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Instance network_info: |[{"id": "29b48485-2c25-460b-a265-d26ccf3a337c", "address": "fa:16:3e:26:c6:46", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29b48485-2c", "ovs_interfaceid": "29b48485-2c25-460b-a265-d26ccf3a337c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1009.682526] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:c6:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '29b48485-2c25-460b-a265-d26ccf3a337c', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.691632] env[63355]: DEBUG oslo.service.loopingcall [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.691988] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1009.692225] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-93036934-2121-4b75-8880-06c9104b2eea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.716470] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.716470] env[63355]: value = "task-1350002" [ 1009.716470] env[63355]: _type = "Task" [ 1009.716470] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.725112] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350002, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.746650] env[63355]: DEBUG nova.compute.manager [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1009.775154] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.775433] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.775602] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.775789] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.775997] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.776141] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.776499] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.776736] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.776980] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.777226] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.777425] env[63355]: DEBUG nova.virt.hardware [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.778363] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121ccf80-a1cc-4975-8887-26b6222cd86e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.789597] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7afdd032-43d7-4b56-9408-c7d48a3a76fa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.795130] env[63355]: DEBUG nova.compute.manager [req-8ff754e9-7067-452a-b85f-006cb8448c80 req-9caaa0bc-8ef7-458d-907b-19ae154bcf75 service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Received event network-changed-29b48485-2c25-460b-a265-d26ccf3a337c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.795404] env[63355]: DEBUG nova.compute.manager [req-8ff754e9-7067-452a-b85f-006cb8448c80 req-9caaa0bc-8ef7-458d-907b-19ae154bcf75 service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Refreshing instance network info cache due to event network-changed-29b48485-2c25-460b-a265-d26ccf3a337c. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1009.795651] env[63355]: DEBUG oslo_concurrency.lockutils [req-8ff754e9-7067-452a-b85f-006cb8448c80 req-9caaa0bc-8ef7-458d-907b-19ae154bcf75 service nova] Acquiring lock "refresh_cache-cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.795815] env[63355]: DEBUG oslo_concurrency.lockutils [req-8ff754e9-7067-452a-b85f-006cb8448c80 req-9caaa0bc-8ef7-458d-907b-19ae154bcf75 service nova] Acquired lock "refresh_cache-cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.795991] env[63355]: DEBUG nova.network.neutron [req-8ff754e9-7067-452a-b85f-006cb8448c80 req-9caaa0bc-8ef7-458d-907b-19ae154bcf75 service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Refreshing network info cache for port 29b48485-2c25-460b-a265-d26ccf3a337c {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.961181] env[63355]: DEBUG oslo_vmware.api [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Task: {'id': task-1350001, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130672} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.961485] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.961700] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.961937] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.962201] env[63355]: INFO nova.compute.manager [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1009.962458] env[63355]: DEBUG oslo.service.loopingcall [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.962682] env[63355]: DEBUG nova.compute.manager [-] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1009.962796] env[63355]: DEBUG nova.network.neutron [-] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1010.074402] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.362s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.074402] env[63355]: INFO nova.compute.manager [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Migrating [ 1010.074402] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.074402] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.075219] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.123s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.077276] env[63355]: INFO nova.compute.claims [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.083020] env[63355]: INFO nova.compute.rpcapi [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 1010.083020] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.100487] env[63355]: DEBUG nova.network.neutron [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Successfully updated port: a0dfd814-02e8-40d1-acbc-d9447b89a9d6 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1010.230074] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350002, 'name': CreateVM_Task, 'duration_secs': 0.412996} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.230304] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1010.231371] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.231563] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.232111] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.232686] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95fc3cba-e8f3-4eb6-823d-9294c06c15a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.237821] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1010.237821] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524b0f0d-8623-0d93-8983-993ce31d39e4" [ 1010.237821] env[63355]: _type = "Task" [ 1010.237821] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.248673] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524b0f0d-8623-0d93-8983-993ce31d39e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.314067] env[63355]: DEBUG nova.compute.manager [req-dfe8291f-39a1-4717-8b03-93d9caf7d12a req-b7f0d83e-038e-4a49-b759-419eb2b17ed0 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Received event network-vif-deleted-40f9f71e-902f-4420-a8b7-58e894a6939c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.314067] env[63355]: INFO nova.compute.manager [req-dfe8291f-39a1-4717-8b03-93d9caf7d12a req-b7f0d83e-038e-4a49-b759-419eb2b17ed0 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Neutron deleted interface 40f9f71e-902f-4420-a8b7-58e894a6939c; detaching it from the instance and deleting it from the info cache [ 1010.314204] env[63355]: DEBUG nova.network.neutron [req-dfe8291f-39a1-4717-8b03-93d9caf7d12a req-b7f0d83e-038e-4a49-b759-419eb2b17ed0 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.599652] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.600041] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.600041] env[63355]: DEBUG nova.network.neutron [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1010.607028] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "refresh_cache-40c48196-6543-4c96-9b8c-3b3c9fbccfe4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.607028] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "refresh_cache-40c48196-6543-4c96-9b8c-3b3c9fbccfe4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.607028] env[63355]: DEBUG nova.network.neutron [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1010.622503] env[63355]: DEBUG nova.network.neutron [req-8ff754e9-7067-452a-b85f-006cb8448c80 req-9caaa0bc-8ef7-458d-907b-19ae154bcf75 service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Updated VIF entry in instance network info cache for port 29b48485-2c25-460b-a265-d26ccf3a337c. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1010.622853] env[63355]: DEBUG nova.network.neutron [req-8ff754e9-7067-452a-b85f-006cb8448c80 req-9caaa0bc-8ef7-458d-907b-19ae154bcf75 service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Updating instance_info_cache with network_info: [{"id": "29b48485-2c25-460b-a265-d26ccf3a337c", "address": "fa:16:3e:26:c6:46", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29b48485-2c", "ovs_interfaceid": "29b48485-2c25-460b-a265-d26ccf3a337c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.749743] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524b0f0d-8623-0d93-8983-993ce31d39e4, 'name': SearchDatastore_Task, 'duration_secs': 0.011468} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.750096] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.750340] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.750607] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.750725] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.751234] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.751524] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9659b797-3a93-42f9-bc34-16eda52fc1f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.761445] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.761764] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1010.762449] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b271263-3bdd-41fb-b179-b4d6a76407a9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.770329] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1010.770329] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52678fbf-0b0a-af4e-757b-11cb69007d7a" [ 1010.770329] env[63355]: _type = "Task" [ 1010.770329] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.779687] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52678fbf-0b0a-af4e-757b-11cb69007d7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.795096] env[63355]: DEBUG nova.network.neutron [-] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.817825] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a545078b-2555-43b1-9837-b2eb7d9c4da4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.828999] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f325ee80-690e-4e5d-b49e-f6d6a001d6de {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.858812] env[63355]: DEBUG nova.compute.manager [req-dfe8291f-39a1-4717-8b03-93d9caf7d12a req-b7f0d83e-038e-4a49-b759-419eb2b17ed0 service nova] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Detach interface failed, port_id=40f9f71e-902f-4420-a8b7-58e894a6939c, reason: Instance 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1011.126226] env[63355]: DEBUG oslo_concurrency.lockutils [req-8ff754e9-7067-452a-b85f-006cb8448c80 req-9caaa0bc-8ef7-458d-907b-19ae154bcf75 service nova] Releasing lock "refresh_cache-cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.139620] env[63355]: DEBUG nova.network.neutron [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1011.313772] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52678fbf-0b0a-af4e-757b-11cb69007d7a, 'name': SearchDatastore_Task, 'duration_secs': 0.01326} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.313772] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a56b9b9-40f2-44d5-b1ea-f60897896331 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.313772] env[63355]: INFO nova.compute.manager [-] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Took 1.34 seconds to deallocate network for instance. [ 1011.313772] env[63355]: DEBUG nova.network.neutron [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Updating instance_info_cache with network_info: [{"id": "a0dfd814-02e8-40d1-acbc-d9447b89a9d6", "address": "fa:16:3e:2c:5e:dd", "network": {"id": "60046874-e6ff-46c0-9843-752048f98e4a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1700834300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b6963b8e16b4986a4545914b75a38ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0dfd814-02", "ovs_interfaceid": "a0dfd814-02e8-40d1-acbc-d9447b89a9d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.313772] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1011.313772] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529d423d-d214-cb3f-fcfc-9ac13ef887ac" [ 1011.313772] env[63355]: _type = "Task" [ 1011.313772] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.328613] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529d423d-d214-cb3f-fcfc-9ac13ef887ac, 'name': SearchDatastore_Task, 'duration_secs': 0.011833} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.328898] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.329282] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] cf2d004b-b9dd-4c05-b54d-9509cd1d0b06/cf2d004b-b9dd-4c05-b54d-9509cd1d0b06.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1011.329566] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8dd367ac-ec73-4efd-b049-250c09e7e1e5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.338025] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54e664c-0144-46a9-95e6-35b24f094a2c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.341915] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1011.341915] env[63355]: value = "task-1350003" [ 1011.341915] env[63355]: _type = "Task" [ 1011.341915] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.349657] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b733f1b5-d03a-45fd-a075-b9e20b9978c0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.356646] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350003, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.386668] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a92607-9f35-458f-9b29-acf52ceb1abc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.393042] env[63355]: DEBUG nova.network.neutron [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance_info_cache with network_info: [{"id": "52d11eb4-329c-4989-9498-377fd46350cb", "address": "fa:16:3e:64:c5:93", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52d11eb4-32", "ovs_interfaceid": "52d11eb4-329c-4989-9498-377fd46350cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.397944] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a477cc9-7a6c-4b3e-a99d-87dfcba1e9f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.414688] env[63355]: DEBUG nova.compute.provider_tree [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.808945] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "refresh_cache-40c48196-6543-4c96-9b8c-3b3c9fbccfe4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.809360] env[63355]: DEBUG nova.compute.manager [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Instance network_info: |[{"id": "a0dfd814-02e8-40d1-acbc-d9447b89a9d6", "address": "fa:16:3e:2c:5e:dd", "network": {"id": "60046874-e6ff-46c0-9843-752048f98e4a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1700834300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b6963b8e16b4986a4545914b75a38ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0dfd814-02", "ovs_interfaceid": "a0dfd814-02e8-40d1-acbc-d9447b89a9d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1011.809926] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:5e:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'df1bf911-aac9-4d2d-ae69-66ace3e6a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0dfd814-02e8-40d1-acbc-d9447b89a9d6', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1011.818324] env[63355]: DEBUG oslo.service.loopingcall [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.819971] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1011.821010] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.822299] env[63355]: DEBUG nova.compute.manager [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Received event network-vif-plugged-a0dfd814-02e8-40d1-acbc-d9447b89a9d6 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.822537] env[63355]: DEBUG oslo_concurrency.lockutils [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] Acquiring lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.822757] env[63355]: DEBUG oslo_concurrency.lockutils [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] Lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.822930] env[63355]: DEBUG oslo_concurrency.lockutils [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] Lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.823151] env[63355]: DEBUG nova.compute.manager [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] No waiting events found dispatching network-vif-plugged-a0dfd814-02e8-40d1-acbc-d9447b89a9d6 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.823329] env[63355]: WARNING nova.compute.manager [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Received unexpected event network-vif-plugged-a0dfd814-02e8-40d1-acbc-d9447b89a9d6 for instance with vm_state building and task_state spawning. [ 1011.823535] env[63355]: DEBUG nova.compute.manager [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Received event network-changed-a0dfd814-02e8-40d1-acbc-d9447b89a9d6 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.823759] env[63355]: DEBUG nova.compute.manager [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Refreshing instance network info cache due to event network-changed-a0dfd814-02e8-40d1-acbc-d9447b89a9d6. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1011.824078] env[63355]: DEBUG oslo_concurrency.lockutils [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] Acquiring lock "refresh_cache-40c48196-6543-4c96-9b8c-3b3c9fbccfe4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.824249] env[63355]: DEBUG oslo_concurrency.lockutils [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] Acquired lock "refresh_cache-40c48196-6543-4c96-9b8c-3b3c9fbccfe4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.824420] env[63355]: DEBUG nova.network.neutron [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Refreshing network info cache for port a0dfd814-02e8-40d1-acbc-d9447b89a9d6 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.825722] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57d31e18-57f6-4e44-97a0-f3446765f915 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.850162] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1011.850162] env[63355]: value = "task-1350004" [ 1011.850162] env[63355]: _type = "Task" [ 1011.850162] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.860930] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350003, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.865196] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350004, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.898241] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.917853] env[63355]: DEBUG nova.scheduler.client.report [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.358120] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350003, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549472} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.358120] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] cf2d004b-b9dd-4c05-b54d-9509cd1d0b06/cf2d004b-b9dd-4c05-b54d-9509cd1d0b06.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1012.358437] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.358596] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa35d842-3de0-4fe4-8904-8bb5b2986b07 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.363474] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350004, 'name': CreateVM_Task, 'duration_secs': 0.435416} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.363931] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1012.364591] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.364757] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.365099] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1012.365345] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f66d313-85ee-4b8a-990a-fb309277cc12 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.369203] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1012.369203] env[63355]: value = "task-1350005" [ 1012.369203] env[63355]: _type = "Task" [ 1012.369203] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.374450] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1012.374450] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b11a1f-bc91-1eaa-e45c-fbeb67fc1c10" [ 1012.374450] env[63355]: _type = "Task" [ 1012.374450] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.382902] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350005, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.388925] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b11a1f-bc91-1eaa-e45c-fbeb67fc1c10, 'name': SearchDatastore_Task, 'duration_secs': 0.011583} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.389251] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.389498] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.389736] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.389886] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.390080] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.390354] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1964681b-d1a3-4c4a-bc4a-c4b7fb49fb3e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.398782] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.398966] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1012.399729] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-279b0030-45b4-4628-9b2a-219bf8613525 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.409444] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1012.409444] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d2d05b-e398-0156-c92c-98445573db74" [ 1012.409444] env[63355]: _type = "Task" [ 1012.409444] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.417751] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d2d05b-e398-0156-c92c-98445573db74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.423950] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.349s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.424611] env[63355]: DEBUG nova.compute.manager [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1012.427173] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.231s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.428650] env[63355]: INFO nova.compute.claims [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.664785] env[63355]: DEBUG nova.network.neutron [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Updated VIF entry in instance network info cache for port a0dfd814-02e8-40d1-acbc-d9447b89a9d6. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1012.665156] env[63355]: DEBUG nova.network.neutron [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Updating instance_info_cache with network_info: [{"id": "a0dfd814-02e8-40d1-acbc-d9447b89a9d6", "address": "fa:16:3e:2c:5e:dd", "network": {"id": "60046874-e6ff-46c0-9843-752048f98e4a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1700834300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b6963b8e16b4986a4545914b75a38ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0dfd814-02", "ovs_interfaceid": "a0dfd814-02e8-40d1-acbc-d9447b89a9d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.880295] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350005, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.175252} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.880295] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1012.881077] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8fbefe4-5cb5-46c9-b60a-fa566e64953b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.903627] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] cf2d004b-b9dd-4c05-b54d-9509cd1d0b06/cf2d004b-b9dd-4c05-b54d-9509cd1d0b06.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.903931] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c36461aa-6f59-4a9d-962c-d6ba3148f773 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.933273] env[63355]: DEBUG nova.compute.utils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1012.937055] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d2d05b-e398-0156-c92c-98445573db74, 'name': SearchDatastore_Task, 'duration_secs': 0.010879} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.938792] env[63355]: DEBUG nova.compute.manager [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1012.939051] env[63355]: DEBUG nova.network.neutron [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1012.941709] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1012.941709] env[63355]: value = "task-1350006" [ 1012.941709] env[63355]: _type = "Task" [ 1012.941709] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.941918] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ea3bf2b-b2ec-47b3-b41e-a443639f203a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.956852] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350006, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.957511] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1012.957511] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524e2ae6-c509-3f5c-9ca8-19bd9e084eef" [ 1012.957511] env[63355]: _type = "Task" [ 1012.957511] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.967665] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524e2ae6-c509-3f5c-9ca8-19bd9e084eef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.987670] env[63355]: DEBUG nova.policy [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'beb41983e69a44b2a108f552da5ef640', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfff1b9903264e5586119ebd3a3602de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1013.168580] env[63355]: DEBUG oslo_concurrency.lockutils [req-a1c6aca9-e71a-42d8-bfa8-f4193421cebe req-e551a851-128c-4d22-a0f7-174b14ff85cc service nova] Releasing lock "refresh_cache-40c48196-6543-4c96-9b8c-3b3c9fbccfe4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.365665] env[63355]: DEBUG nova.network.neutron [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Successfully created port: 6d3a3c43-ae33-47e1-9e1b-bf0d9750e565 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1013.425620] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffb23cf-894a-47f1-8fca-8639f9b6b360 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.447194] env[63355]: DEBUG nova.compute.manager [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1013.454160] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance 'fd55ddfc-f376-4e15-961e-6826f31a7890' progress to 0 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1013.472695] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524e2ae6-c509-3f5c-9ca8-19bd9e084eef, 'name': SearchDatastore_Task, 'duration_secs': 0.013091} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.476370] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.476791] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 40c48196-6543-4c96-9b8c-3b3c9fbccfe4/40c48196-6543-4c96-9b8c-3b3c9fbccfe4.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1013.477169] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350006, 'name': ReconfigVM_Task, 'duration_secs': 0.313364} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.480398] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a80d269c-a4a8-4f13-b3de-398741da21e7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.482476] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Reconfigured VM instance instance-0000005a to attach disk [datastore1] cf2d004b-b9dd-4c05-b54d-9509cd1d0b06/cf2d004b-b9dd-4c05-b54d-9509cd1d0b06.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.483407] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d828a91-ba3b-463d-996b-ba467c042b96 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.492098] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1013.492098] env[63355]: value = "task-1350007" [ 1013.492098] env[63355]: _type = "Task" [ 1013.492098] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.493619] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1013.493619] env[63355]: value = "task-1350008" [ 1013.493619] env[63355]: _type = "Task" [ 1013.493619] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.513307] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350007, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.681235] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30394b71-6480-4f20-b8ad-63ae8fc0c934 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.691028] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c282112a-cf94-4dea-b03d-1fae2e882b2b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.723908] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c665fde-0e4d-49fb-bba8-b407be3a5ce9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.733539] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648ebe69-453c-4d09-bd87-b50ac38859cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.752254] env[63355]: DEBUG nova.compute.provider_tree [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.960647] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1013.960647] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c44a45c0-6e69-4776-8f07-049191e4b7ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.971130] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1013.971130] env[63355]: value = "task-1350009" [ 1013.971130] env[63355]: _type = "Task" [ 1013.971130] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.982941] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350009, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.013332] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350007, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.018075] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350008, 'name': Rename_Task, 'duration_secs': 0.175172} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.018485] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1014.018991] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d729997-2c9f-47c2-a4e0-b2c8d35d1780 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.029096] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1014.029096] env[63355]: value = "task-1350010" [ 1014.029096] env[63355]: _type = "Task" [ 1014.029096] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.039801] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350010, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.260404] env[63355]: DEBUG nova.scheduler.client.report [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.460186] env[63355]: DEBUG nova.compute.manager [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1014.483451] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350009, 'name': PowerOffVM_Task, 'duration_secs': 0.252968} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.485828] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.486081] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.486269] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.486523] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.486791] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.486866] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.487389] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.487389] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.487494] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.487634] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.487810] env[63355]: DEBUG nova.virt.hardware [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.488524] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.488524] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance 'fd55ddfc-f376-4e15-961e-6826f31a7890' progress to 17 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1014.492665] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c6d271-e711-4d41-86ba-8afacafab384 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.503813] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350007, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568425} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.505954] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 40c48196-6543-4c96-9b8c-3b3c9fbccfe4/40c48196-6543-4c96-9b8c-3b3c9fbccfe4.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1014.506222] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1014.506554] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c347007d-4331-4410-866b-8e73cb37c012 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.509315] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71188555-55d2-4275-8e8d-f7dcaca7ef26 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.527222] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1014.527222] env[63355]: value = "task-1350011" [ 1014.527222] env[63355]: _type = "Task" [ 1014.527222] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.540700] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350011, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.543843] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350010, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.573270] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.573590] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.765538] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.766189] env[63355]: DEBUG nova.compute.manager [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1014.769732] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.196s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.769732] env[63355]: DEBUG nova.objects.instance [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lazy-loading 'resources' on Instance uuid 8e3ec9d3-bc22-4e39-ad7c-93268dd59020 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.980567] env[63355]: DEBUG nova.network.neutron [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Successfully updated port: 6d3a3c43-ae33-47e1-9e1b-bf0d9750e565 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1014.983319] env[63355]: DEBUG nova.compute.manager [req-c31d3f8f-e255-440c-901b-85503c29003e req-5a3f18ed-ed59-4d4e-b104-31b5a2439bb0 service nova] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Received event network-vif-plugged-6d3a3c43-ae33-47e1-9e1b-bf0d9750e565 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.983578] env[63355]: DEBUG oslo_concurrency.lockutils [req-c31d3f8f-e255-440c-901b-85503c29003e req-5a3f18ed-ed59-4d4e-b104-31b5a2439bb0 service nova] Acquiring lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.983752] env[63355]: DEBUG oslo_concurrency.lockutils [req-c31d3f8f-e255-440c-901b-85503c29003e req-5a3f18ed-ed59-4d4e-b104-31b5a2439bb0 service nova] Lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.983943] env[63355]: DEBUG oslo_concurrency.lockutils [req-c31d3f8f-e255-440c-901b-85503c29003e req-5a3f18ed-ed59-4d4e-b104-31b5a2439bb0 service nova] Lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.984596] env[63355]: DEBUG nova.compute.manager [req-c31d3f8f-e255-440c-901b-85503c29003e req-5a3f18ed-ed59-4d4e-b104-31b5a2439bb0 service nova] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] No waiting events found dispatching network-vif-plugged-6d3a3c43-ae33-47e1-9e1b-bf0d9750e565 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.984596] env[63355]: WARNING nova.compute.manager [req-c31d3f8f-e255-440c-901b-85503c29003e req-5a3f18ed-ed59-4d4e-b104-31b5a2439bb0 service nova] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Received unexpected event network-vif-plugged-6d3a3c43-ae33-47e1-9e1b-bf0d9750e565 for instance with vm_state building and task_state spawning. [ 1014.997881] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.999298] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.999298] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.999298] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.999298] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.999495] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.000260] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.000260] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.000478] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.000692] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.001132] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.007591] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46441c0d-5e44-405d-98a7-5f167d65d6a4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.026025] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1015.026025] env[63355]: value = "task-1350012" [ 1015.026025] env[63355]: _type = "Task" [ 1015.026025] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.040313] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350012, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.044591] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350011, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.205362} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.045159] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1015.046363] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7ce59e-ea1d-4299-b73c-2ea71aef3a90 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.053606] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350010, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.080337] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 40c48196-6543-4c96-9b8c-3b3c9fbccfe4/40c48196-6543-4c96-9b8c-3b3c9fbccfe4.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.081764] env[63355]: DEBUG nova.compute.utils [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1015.083524] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c503f00f-7612-4726-86ae-f380810069d7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.106369] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1015.106369] env[63355]: value = "task-1350013" [ 1015.106369] env[63355]: _type = "Task" [ 1015.106369] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.115596] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350013, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.272605] env[63355]: DEBUG nova.compute.utils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1015.281119] env[63355]: DEBUG nova.compute.manager [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1015.281509] env[63355]: DEBUG nova.network.neutron [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1015.351973] env[63355]: DEBUG nova.policy [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77243643ea724b72858a8682a2a054a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d47b671ea9c429391cbdae7e24adadf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1015.467929] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e2ac0bd-bdf5-47b5-93e8-d0c76a73ae24 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.475898] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4caad4-2f72-4f19-b922-115e5ccca697 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.507619] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.507899] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.507992] env[63355]: DEBUG nova.network.neutron [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1015.510597] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d162459-5ad2-4e00-871b-94cc7ff26697 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.519128] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad68f462-3d99-4372-99e3-52aa1659c4c8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.533676] env[63355]: DEBUG nova.compute.provider_tree [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.547694] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350012, 'name': ReconfigVM_Task, 'duration_secs': 0.252978} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.551047] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance 'fd55ddfc-f376-4e15-961e-6826f31a7890' progress to 33 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1015.555063] env[63355]: DEBUG oslo_vmware.api [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350010, 'name': PowerOnVM_Task, 'duration_secs': 1.045815} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.556092] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1015.556201] env[63355]: INFO nova.compute.manager [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Took 8.13 seconds to spawn the instance on the hypervisor. [ 1015.556453] env[63355]: DEBUG nova.compute.manager [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.557273] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024fa989-4be2-425f-96db-3c147c0b281e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.585532] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.619979] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350013, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.783439] env[63355]: DEBUG nova.compute.manager [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1015.973132] env[63355]: DEBUG nova.network.neutron [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Successfully created port: 546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1016.037203] env[63355]: DEBUG nova.scheduler.client.report [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.045262] env[63355]: DEBUG nova.network.neutron [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1016.057792] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.058134] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.058364] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.059055] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.059055] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.059055] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.059320] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.059320] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.059574] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.059728] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.059931] env[63355]: DEBUG nova.virt.hardware [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.065766] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Reconfiguring VM instance instance-00000055 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1016.069120] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21d354eb-505c-457f-97d4-1d25cead437c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.089786] env[63355]: INFO nova.compute.manager [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Took 15.21 seconds to build instance. [ 1016.099280] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1016.099280] env[63355]: value = "task-1350014" [ 1016.099280] env[63355]: _type = "Task" [ 1016.099280] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.109506] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350014, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.121878] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350013, 'name': ReconfigVM_Task, 'duration_secs': 0.549087} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.122215] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 40c48196-6543-4c96-9b8c-3b3c9fbccfe4/40c48196-6543-4c96-9b8c-3b3c9fbccfe4.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.122970] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b974667-5f1d-44f5-8b0a-6211b4bc74ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.132301] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1016.132301] env[63355]: value = "task-1350015" [ 1016.132301] env[63355]: _type = "Task" [ 1016.132301] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.142885] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350015, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.231916] env[63355]: DEBUG nova.network.neutron [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Updating instance_info_cache with network_info: [{"id": "6d3a3c43-ae33-47e1-9e1b-bf0d9750e565", "address": "fa:16:3e:43:6b:bc", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d3a3c43-ae", "ovs_interfaceid": "6d3a3c43-ae33-47e1-9e1b-bf0d9750e565", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.476336] env[63355]: DEBUG oslo_vmware.rw_handles [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5a3e6-f9e5-7f87-64cf-90f880ce5276/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1016.476336] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3183fb41-87a9-49a3-950e-82110428fddd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.483160] env[63355]: DEBUG oslo_vmware.rw_handles [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5a3e6-f9e5-7f87-64cf-90f880ce5276/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1016.483421] env[63355]: ERROR oslo_vmware.rw_handles [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5a3e6-f9e5-7f87-64cf-90f880ce5276/disk-0.vmdk due to incomplete transfer. [ 1016.483724] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-60afbaa1-82ba-4cad-abca-d37f87833c47 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.491958] env[63355]: DEBUG oslo_vmware.rw_handles [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5a3e6-f9e5-7f87-64cf-90f880ce5276/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1016.492193] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Uploaded image 3061ee1f-5dc1-4bba-9208-313cf0785cd6 to the Glance image server {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1016.494463] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1016.496173] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-639e5162-13db-429a-8060-72373d2bfcf4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.505514] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1016.505514] env[63355]: value = "task-1350016" [ 1016.505514] env[63355]: _type = "Task" [ 1016.505514] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.514303] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350016, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.542772] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.544920] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.684s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.545204] env[63355]: DEBUG nova.objects.instance [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lazy-loading 'resources' on Instance uuid 7c153109-b814-4e11-b4f9-7b8cebb853d4 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.567396] env[63355]: INFO nova.scheduler.client.report [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Deleted allocations for instance 8e3ec9d3-bc22-4e39-ad7c-93268dd59020 [ 1016.593092] env[63355]: DEBUG oslo_concurrency.lockutils [None req-faa71d5c-2fa0-4dbc-93a0-d7333b5e603d tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.719s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.614153] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350014, 'name': ReconfigVM_Task, 'duration_secs': 0.181114} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.614153] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Reconfigured VM instance instance-00000055 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1016.614800] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4148774d-0ca1-4d7f-87e1-9f49ba0aab11 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.639989] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] fd55ddfc-f376-4e15-961e-6826f31a7890/fd55ddfc-f376-4e15-961e-6826f31a7890.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.640692] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4570a042-948b-4444-9d17-13e6d44c90c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.666218] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350015, 'name': Rename_Task, 'duration_secs': 0.231002} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.667760] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1016.668107] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1016.668107] env[63355]: value = "task-1350017" [ 1016.668107] env[63355]: _type = "Task" [ 1016.668107] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.668327] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0846cbde-cda0-4abe-b6ab-92a445a35192 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.682418] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.682900] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.683339] env[63355]: INFO nova.compute.manager [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Attaching volume 64055852-0585-4516-8f95-0996e47ccc04 to /dev/sdb [ 1016.686080] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350017, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.689051] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1016.689051] env[63355]: value = "task-1350018" [ 1016.689051] env[63355]: _type = "Task" [ 1016.689051] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.707113] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350018, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.726402] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e31097a-1f20-4ee7-a935-ae2aa2b05ee4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.735095] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.735444] env[63355]: DEBUG nova.compute.manager [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Instance network_info: |[{"id": "6d3a3c43-ae33-47e1-9e1b-bf0d9750e565", "address": "fa:16:3e:43:6b:bc", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d3a3c43-ae", "ovs_interfaceid": "6d3a3c43-ae33-47e1-9e1b-bf0d9750e565", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1016.735961] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:6b:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '52358fcc-0d9f-45dd-8c75-db533fd992c3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d3a3c43-ae33-47e1-9e1b-bf0d9750e565', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1016.743892] env[63355]: DEBUG oslo.service.loopingcall [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.744651] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70dca3b3-218c-4dcb-9619-03f0be7f05d2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.747685] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1016.747934] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3adf474c-bbd7-4dc5-92a7-7381aab4608e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.772843] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1016.772843] env[63355]: value = "task-1350019" [ 1016.772843] env[63355]: _type = "Task" [ 1016.772843] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.780941] env[63355]: DEBUG nova.virt.block_device [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updating existing volume attachment record: 81418b9e-e22a-4d65-a5de-6c4fb38dc463 {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1016.789398] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350019, 'name': CreateVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.797056] env[63355]: DEBUG nova.compute.manager [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1016.824839] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.825232] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.828077] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.828077] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.828077] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.828077] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.828077] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.828077] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.828077] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.828077] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.828077] env[63355]: DEBUG nova.virt.hardware [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.828077] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9545a7ea-ad10-4aa0-b27c-e625e24b20aa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.837440] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9cbc58-15c5-4888-8df7-9a0ee46bbf44 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.013475] env[63355]: DEBUG nova.compute.manager [req-d89ed534-0d00-49f2-b887-096868851071 req-942c66af-617e-43e3-b472-40d25d0cb962 service nova] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Received event network-changed-6d3a3c43-ae33-47e1-9e1b-bf0d9750e565 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.013680] env[63355]: DEBUG nova.compute.manager [req-d89ed534-0d00-49f2-b887-096868851071 req-942c66af-617e-43e3-b472-40d25d0cb962 service nova] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Refreshing instance network info cache due to event network-changed-6d3a3c43-ae33-47e1-9e1b-bf0d9750e565. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1017.013900] env[63355]: DEBUG oslo_concurrency.lockutils [req-d89ed534-0d00-49f2-b887-096868851071 req-942c66af-617e-43e3-b472-40d25d0cb962 service nova] Acquiring lock "refresh_cache-2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.014060] env[63355]: DEBUG oslo_concurrency.lockutils [req-d89ed534-0d00-49f2-b887-096868851071 req-942c66af-617e-43e3-b472-40d25d0cb962 service nova] Acquired lock "refresh_cache-2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.014233] env[63355]: DEBUG nova.network.neutron [req-d89ed534-0d00-49f2-b887-096868851071 req-942c66af-617e-43e3-b472-40d25d0cb962 service nova] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Refreshing network info cache for port 6d3a3c43-ae33-47e1-9e1b-bf0d9750e565 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1017.021044] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350016, 'name': Destroy_Task, 'duration_secs': 0.386267} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.021615] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Destroyed the VM [ 1017.021940] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1017.022220] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9ec6a7dc-d245-4f30-b9f0-701e271cfff9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.030521] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1017.030521] env[63355]: value = "task-1350022" [ 1017.030521] env[63355]: _type = "Task" [ 1017.030521] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.044108] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350022, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.047558] env[63355]: DEBUG nova.objects.instance [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lazy-loading 'numa_topology' on Instance uuid 7c153109-b814-4e11-b4f9-7b8cebb853d4 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.076538] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8f5f4e6d-fb20-4414-8bd8-50e2d91a439a tempest-ServerRescueTestJSON-1144335544 tempest-ServerRescueTestJSON-1144335544-project-member] Lock "8e3ec9d3-bc22-4e39-ad7c-93268dd59020" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.722s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.180461] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350017, 'name': ReconfigVM_Task, 'duration_secs': 0.307233} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.180798] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Reconfigured VM instance instance-00000055 to attach disk [datastore2] fd55ddfc-f376-4e15-961e-6826f31a7890/fd55ddfc-f376-4e15-961e-6826f31a7890.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.181130] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance 'fd55ddfc-f376-4e15-961e-6826f31a7890' progress to 50 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1017.198316] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350018, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.209050] env[63355]: DEBUG oslo_concurrency.lockutils [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.209217] env[63355]: DEBUG oslo_concurrency.lockutils [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.209401] env[63355]: DEBUG oslo_concurrency.lockutils [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.209585] env[63355]: DEBUG oslo_concurrency.lockutils [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.209753] env[63355]: DEBUG oslo_concurrency.lockutils [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.212223] env[63355]: INFO nova.compute.manager [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Terminating instance [ 1017.214532] env[63355]: DEBUG nova.compute.manager [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1017.214780] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1017.215623] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2734c647-2cbc-4ba7-81b6-96e30fbae2d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.225771] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1017.226081] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-184e75df-1119-45cf-9e0e-7385173553b6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.235961] env[63355]: DEBUG oslo_vmware.api [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1017.235961] env[63355]: value = "task-1350024" [ 1017.235961] env[63355]: _type = "Task" [ 1017.235961] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.246066] env[63355]: DEBUG oslo_vmware.api [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350024, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.281849] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350019, 'name': CreateVM_Task, 'duration_secs': 0.454656} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.282068] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1017.282758] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.282932] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.283283] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.283655] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-587a3252-b3a7-4072-9666-74e3a98afe18 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.289239] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1017.289239] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]520c7dc5-b57f-3020-4d85-e4cad5e2dc44" [ 1017.289239] env[63355]: _type = "Task" [ 1017.289239] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.297822] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]520c7dc5-b57f-3020-4d85-e4cad5e2dc44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.544775] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350022, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.550510] env[63355]: DEBUG nova.objects.base [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Object Instance<7c153109-b814-4e11-b4f9-7b8cebb853d4> lazy-loaded attributes: resources,numa_topology {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1017.689902] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7df14b9-f59c-4388-a416-4df89baae4cc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.717716] env[63355]: DEBUG oslo_vmware.api [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350018, 'name': PowerOnVM_Task, 'duration_secs': 0.877774} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.720814] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f293c31d-9372-4f4b-8be2-5c39555abda6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.723320] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1017.723541] env[63355]: INFO nova.compute.manager [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Took 7.98 seconds to spawn the instance on the hypervisor. [ 1017.723725] env[63355]: DEBUG nova.compute.manager [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.724715] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c2bdbc-a88e-4dd4-aaf0-9fbd8c07a344 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.744910] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance 'fd55ddfc-f376-4e15-961e-6826f31a7890' progress to 67 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1017.760023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9b4260-6784-4f10-9e08-d0a84f4efe31 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.760467] env[63355]: DEBUG nova.network.neutron [req-d89ed534-0d00-49f2-b887-096868851071 req-942c66af-617e-43e3-b472-40d25d0cb962 service nova] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Updated VIF entry in instance network info cache for port 6d3a3c43-ae33-47e1-9e1b-bf0d9750e565. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1017.760793] env[63355]: DEBUG nova.network.neutron [req-d89ed534-0d00-49f2-b887-096868851071 req-942c66af-617e-43e3-b472-40d25d0cb962 service nova] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Updating instance_info_cache with network_info: [{"id": "6d3a3c43-ae33-47e1-9e1b-bf0d9750e565", "address": "fa:16:3e:43:6b:bc", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d3a3c43-ae", "ovs_interfaceid": "6d3a3c43-ae33-47e1-9e1b-bf0d9750e565", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.770172] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca489b1-916e-4241-8352-86aee51357ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.773306] env[63355]: DEBUG oslo_vmware.api [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350024, 'name': PowerOffVM_Task, 'duration_secs': 0.200349} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.778021] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1017.778021] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1017.778021] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5aa15e98-f609-489b-bfff-a94dafa2a524 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.806498] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a7b15f-d118-46a3-a7c7-f74aedbae8d1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.818240] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]520c7dc5-b57f-3020-4d85-e4cad5e2dc44, 'name': SearchDatastore_Task, 'duration_secs': 0.016718} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.818648] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.818887] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1017.819141] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.819296] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.819482] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1017.820698] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a4776f-d2be-4d56-a62c-eeac0853a205 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.824432] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de795bb2-ba56-49fb-9ef7-0fa715ecda0a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.837019] env[63355]: DEBUG nova.compute.provider_tree [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.842021] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1017.842021] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1017.842021] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a3e94d2-3a37-4a29-83c0-b954b262a914 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.846498] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1017.846498] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52910c09-5e8c-dddb-967e-42622e058383" [ 1017.846498] env[63355]: _type = "Task" [ 1017.846498] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.857542] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52910c09-5e8c-dddb-967e-42622e058383, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.868518] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1017.868795] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1017.868987] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleting the datastore file [datastore1] cf2d004b-b9dd-4c05-b54d-9509cd1d0b06 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1017.869287] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28849dda-0654-498b-b9d2-a3d612146f0b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.876746] env[63355]: DEBUG oslo_vmware.api [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1017.876746] env[63355]: value = "task-1350026" [ 1017.876746] env[63355]: _type = "Task" [ 1017.876746] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.886243] env[63355]: DEBUG oslo_vmware.api [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350026, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.047055] env[63355]: DEBUG oslo_vmware.api [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350022, 'name': RemoveSnapshot_Task, 'duration_secs': 0.538101} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.047602] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1018.048028] env[63355]: INFO nova.compute.manager [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Took 14.87 seconds to snapshot the instance on the hypervisor. [ 1018.152106] env[63355]: DEBUG nova.network.neutron [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Successfully updated port: 546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.270343] env[63355]: DEBUG oslo_concurrency.lockutils [req-d89ed534-0d00-49f2-b887-096868851071 req-942c66af-617e-43e3-b472-40d25d0cb962 service nova] Releasing lock "refresh_cache-2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.276666] env[63355]: INFO nova.compute.manager [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Took 17.39 seconds to build instance. [ 1018.342193] env[63355]: DEBUG nova.network.neutron [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Port 52d11eb4-329c-4989-9498-377fd46350cb binding to destination host cpu-1 is already ACTIVE {{(pid=63355) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1018.342723] env[63355]: DEBUG nova.scheduler.client.report [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.360034] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52910c09-5e8c-dddb-967e-42622e058383, 'name': SearchDatastore_Task, 'duration_secs': 0.038034} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.361847] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-690b4517-d692-42e2-94c4-ceea8726e2ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.368120] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1018.368120] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5239912a-6f76-2e87-4640-f14afbe2bf05" [ 1018.368120] env[63355]: _type = "Task" [ 1018.368120] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.376956] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5239912a-6f76-2e87-4640-f14afbe2bf05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.384751] env[63355]: DEBUG oslo_vmware.api [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350026, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.298494} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.384996] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1018.385200] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1018.385381] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1018.385558] env[63355]: INFO nova.compute.manager [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1018.385795] env[63355]: DEBUG oslo.service.loopingcall [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.385985] env[63355]: DEBUG nova.compute.manager [-] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1018.386092] env[63355]: DEBUG nova.network.neutron [-] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1018.601702] env[63355]: DEBUG nova.compute.manager [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Found 3 images (rotation: 2) {{(pid=63355) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1018.601702] env[63355]: DEBUG nova.compute.manager [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Rotating out 1 backups {{(pid=63355) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1018.601702] env[63355]: DEBUG nova.compute.manager [None req-09dacda7-a9e7-445f-9e9f-724ba677bab1 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Deleting image eabfdc0f-53a1-4aa1-a863-de4d0008a5e5 {{(pid=63355) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1018.653721] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.653882] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.654052] env[63355]: DEBUG nova.network.neutron [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1018.779184] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d484df88-7b10-480e-afde-7a02ad125d7d tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.903s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.853953] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.309s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.856484] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.101s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.856760] env[63355]: DEBUG nova.objects.instance [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lazy-loading 'resources' on Instance uuid 9ee3001b-8bf0-43ab-996a-a68dad57d8e6 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.881614] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5239912a-6f76-2e87-4640-f14afbe2bf05, 'name': SearchDatastore_Task, 'duration_secs': 0.010041} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.882709] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.882996] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07/2214ffc8-d0fd-49f3-91c1-74c13ef7bc07.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1018.883306] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe277bcf-f1a6-4b94-bc22-965455721f33 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.892287] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1018.892287] env[63355]: value = "task-1350027" [ 1018.892287] env[63355]: _type = "Task" [ 1018.892287] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.908077] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350027, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.105244] env[63355]: DEBUG nova.compute.manager [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Received event network-vif-plugged-546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1019.105554] env[63355]: DEBUG oslo_concurrency.lockutils [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] Acquiring lock "fa748748-31ec-431a-a628-5ea179e26fc8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.109754] env[63355]: DEBUG oslo_concurrency.lockutils [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] Lock "fa748748-31ec-431a-a628-5ea179e26fc8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.109754] env[63355]: DEBUG oslo_concurrency.lockutils [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] Lock "fa748748-31ec-431a-a628-5ea179e26fc8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.109754] env[63355]: DEBUG nova.compute.manager [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] No waiting events found dispatching network-vif-plugged-546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1019.109754] env[63355]: WARNING nova.compute.manager [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Received unexpected event network-vif-plugged-546aac3d-2d71-48bd-84ac-6161e42a90dd for instance with vm_state building and task_state spawning. [ 1019.109754] env[63355]: DEBUG nova.compute.manager [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Received event network-changed-546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1019.109754] env[63355]: DEBUG nova.compute.manager [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Refreshing instance network info cache due to event network-changed-546aac3d-2d71-48bd-84ac-6161e42a90dd. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1019.109754] env[63355]: DEBUG oslo_concurrency.lockutils [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] Acquiring lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.212200] env[63355]: DEBUG nova.network.neutron [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1019.297711] env[63355]: DEBUG nova.network.neutron [-] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.387023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "fd55ddfc-f376-4e15-961e-6826f31a7890-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.387023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.387023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.387023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d6b13e4-e4be-472d-8361-3817d9e7e9e8 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 35.977s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.388141] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 12.183s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.388323] env[63355]: INFO nova.compute.manager [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Unshelving [ 1019.406538] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350027, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504628} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.407773] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07/2214ffc8-d0fd-49f3-91c1-74c13ef7bc07.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1019.408475] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1019.409345] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f64658a0-f5f3-49b8-8fb7-cb791b1b1294 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.419018] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1019.419018] env[63355]: value = "task-1350029" [ 1019.419018] env[63355]: _type = "Task" [ 1019.419018] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.437449] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350029, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.539949] env[63355]: DEBUG nova.network.neutron [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updating instance_info_cache with network_info: [{"id": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "address": "fa:16:3e:40:51:98", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546aac3d-2d", "ovs_interfaceid": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.650628] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66385b46-e871-41c7-b03e-2140092c9ffa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.659886] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2a8383-4820-4a86-a1d1-fce3fa18da51 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.696031] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8780ee78-984a-4077-9089-8e14b703a9b4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.703359] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85aeb6c-6552-4735-a8ac-f9688f16d232 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.720029] env[63355]: DEBUG nova.compute.provider_tree [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.806968] env[63355]: INFO nova.compute.manager [-] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Took 1.42 seconds to deallocate network for instance. [ 1019.931898] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350029, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0702} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.932620] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1019.934306] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9020a16-f272-4776-8156-caa46bc26bee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.957653] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07/2214ffc8-d0fd-49f3-91c1-74c13ef7bc07.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1019.958957] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ec0cf45-7d21-4fc7-af20-28df8209bc3f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.979957] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1019.979957] env[63355]: value = "task-1350030" [ 1019.979957] env[63355]: _type = "Task" [ 1019.979957] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.988468] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350030, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.043892] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.044642] env[63355]: DEBUG nova.compute.manager [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Instance network_info: |[{"id": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "address": "fa:16:3e:40:51:98", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546aac3d-2d", "ovs_interfaceid": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1020.044825] env[63355]: DEBUG oslo_concurrency.lockutils [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] Acquired lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.044967] env[63355]: DEBUG nova.network.neutron [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Refreshing network info cache for port 546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1020.046317] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:51:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e41070eb-3ac1-4ca9-a3d0-fd65893a97de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '546aac3d-2d71-48bd-84ac-6161e42a90dd', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.058271] env[63355]: DEBUG oslo.service.loopingcall [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.061699] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1020.062302] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6de7050e-6590-437e-be25-8953c937bdde {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.093319] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.093319] env[63355]: value = "task-1350031" [ 1020.093319] env[63355]: _type = "Task" [ 1020.093319] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.107714] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350031, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.198059] env[63355]: DEBUG nova.compute.manager [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.199092] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a89670-8e1d-4bb1-908d-8456f1ea31b4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.222632] env[63355]: DEBUG nova.scheduler.client.report [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.314710] env[63355]: DEBUG oslo_concurrency.lockutils [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.441334] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.491704] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.604043] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350031, 'name': CreateVM_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.670553] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.670730] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.670920] env[63355]: DEBUG nova.network.neutron [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1020.706731] env[63355]: DEBUG nova.network.neutron [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updated VIF entry in instance network info cache for port 546aac3d-2d71-48bd-84ac-6161e42a90dd. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1020.707132] env[63355]: DEBUG nova.network.neutron [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updating instance_info_cache with network_info: [{"id": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "address": "fa:16:3e:40:51:98", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546aac3d-2d", "ovs_interfaceid": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.711072] env[63355]: INFO nova.compute.manager [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] instance snapshotting [ 1020.717596] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfb3cd7-99a0-43c5-b827-073811e067df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.737465] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.881s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.739829] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.919s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.740077] env[63355]: DEBUG nova.objects.instance [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lazy-loading 'resources' on Instance uuid 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.741575] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69fcec3-230e-4ef6-be3b-3ee0e53bb0fa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.764173] env[63355]: INFO nova.scheduler.client.report [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Deleted allocations for instance 9ee3001b-8bf0-43ab-996a-a68dad57d8e6 [ 1020.992126] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350030, 'name': ReconfigVM_Task, 'duration_secs': 0.648976} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.992637] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07/2214ffc8-d0fd-49f3-91c1-74c13ef7bc07.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1020.993356] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c536425-f814-4b87-944f-892b81199a6b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.001515] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1021.001515] env[63355]: value = "task-1350032" [ 1021.001515] env[63355]: _type = "Task" [ 1021.001515] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.012920] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350032, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.105216] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350031, 'name': CreateVM_Task, 'duration_secs': 0.866744} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.105282] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1021.105958] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.106151] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.106534] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1021.107179] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f643490-05f7-4bce-a1b6-37e8dbd6864f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.111894] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1021.111894] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5258fa28-ab63-20d4-68d6-74c662fca116" [ 1021.111894] env[63355]: _type = "Task" [ 1021.111894] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.120909] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5258fa28-ab63-20d4-68d6-74c662fca116, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.210797] env[63355]: DEBUG oslo_concurrency.lockutils [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] Releasing lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.210797] env[63355]: DEBUG nova.compute.manager [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Received event network-vif-deleted-29b48485-2c25-460b-a265-d26ccf3a337c {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.211113] env[63355]: INFO nova.compute.manager [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Neutron deleted interface 29b48485-2c25-460b-a265-d26ccf3a337c; detaching it from the instance and deleting it from the info cache [ 1021.211113] env[63355]: DEBUG nova.network.neutron [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.256311] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1021.256653] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-42763601-304e-445c-a326-e468ec33d5c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.265061] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1021.265061] env[63355]: value = "task-1350033" [ 1021.265061] env[63355]: _type = "Task" [ 1021.265061] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.275653] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350033, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.279386] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0a67319-f791-496f-8df2-498991421f65 tempest-ServerMetadataNegativeTestJSON-1683749208 tempest-ServerMetadataNegativeTestJSON-1683749208-project-member] Lock "9ee3001b-8bf0-43ab-996a-a68dad57d8e6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.448s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.342749] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1021.342994] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287741', 'volume_id': '64055852-0585-4516-8f95-0996e47ccc04', 'name': 'volume-64055852-0585-4516-8f95-0996e47ccc04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae3961d2-dc5b-4e49-acca-6fb52291f23b', 'attached_at': '', 'detached_at': '', 'volume_id': '64055852-0585-4516-8f95-0996e47ccc04', 'serial': '64055852-0585-4516-8f95-0996e47ccc04'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1021.344029] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3389f341-9e1e-4fc3-8236-0e880fb4baba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.363811] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b9d30f-e6a9-4325-abe5-f41f67cf24a0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.393226] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] volume-64055852-0585-4516-8f95-0996e47ccc04/volume-64055852-0585-4516-8f95-0996e47ccc04.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.395796] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7f219c1-722b-488d-b659-03baf9915f8f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.415056] env[63355]: DEBUG oslo_vmware.api [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1021.415056] env[63355]: value = "task-1350034" [ 1021.415056] env[63355]: _type = "Task" [ 1021.415056] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.425948] env[63355]: DEBUG oslo_vmware.api [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350034, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.493440] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc5b576-6dde-40a2-8077-95fbdad246d5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.507906] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86880958-1788-443d-942b-43da8de90d6f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.542715] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d044be7-c29d-4355-9fc6-5d6705ce2501 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.546106] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350032, 'name': Rename_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.552203] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a692f1d-51e2-45c6-97c8-9bd89c8e4c81 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.568485] env[63355]: DEBUG nova.compute.provider_tree [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.581044] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.581326] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.626524] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5258fa28-ab63-20d4-68d6-74c662fca116, 'name': SearchDatastore_Task, 'duration_secs': 0.02063} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.626844] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.627103] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.627348] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.627553] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.627728] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.627991] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-179f86e4-23a2-4228-94cc-b5ad769ecdf1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.636551] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.636737] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.637452] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a36a85d2-d333-4da5-944e-e6135a9beb51 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.643106] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1021.643106] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fb6fae-7afc-1a9c-598f-a9422b3e31d5" [ 1021.643106] env[63355]: _type = "Task" [ 1021.643106] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.648344] env[63355]: DEBUG nova.network.neutron [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance_info_cache with network_info: [{"id": "52d11eb4-329c-4989-9498-377fd46350cb", "address": "fa:16:3e:64:c5:93", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52d11eb4-32", "ovs_interfaceid": "52d11eb4-329c-4989-9498-377fd46350cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.652354] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fb6fae-7afc-1a9c-598f-a9422b3e31d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.715045] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63e1b5b8-34da-4129-bc3d-34e5e6888166 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.724514] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d71c7c-6cfb-4ca8-9c08-2b10b0ee7793 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.758095] env[63355]: DEBUG nova.compute.manager [req-064e0431-db79-4e6b-bd8a-001104e73e50 req-6a2b1494-4d01-49c6-b105-c204b28429bc service nova] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Detach interface failed, port_id=29b48485-2c25-460b-a265-d26ccf3a337c, reason: Instance cf2d004b-b9dd-4c05-b54d-9509cd1d0b06 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1021.773915] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350033, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.925444] env[63355]: DEBUG oslo_vmware.api [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350034, 'name': ReconfigVM_Task, 'duration_secs': 0.370238} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.925730] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Reconfigured VM instance instance-00000053 to attach disk [datastore2] volume-64055852-0585-4516-8f95-0996e47ccc04/volume-64055852-0585-4516-8f95-0996e47ccc04.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.930702] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb42c573-a7d3-4b41-a3fb-498f8001a3c0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.947362] env[63355]: DEBUG oslo_vmware.api [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1021.947362] env[63355]: value = "task-1350035" [ 1021.947362] env[63355]: _type = "Task" [ 1021.947362] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.957429] env[63355]: DEBUG oslo_vmware.api [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350035, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.015401] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350032, 'name': Rename_Task, 'duration_secs': 0.971627} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.015697] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.015950] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5184e405-1e64-497a-acff-c31f5961ea47 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.026767] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1022.026767] env[63355]: value = "task-1350036" [ 1022.026767] env[63355]: _type = "Task" [ 1022.026767] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.035228] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350036, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.073595] env[63355]: DEBUG nova.scheduler.client.report [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.077652] env[63355]: DEBUG oslo_concurrency.lockutils [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.077909] env[63355]: DEBUG oslo_concurrency.lockutils [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.078102] env[63355]: DEBUG nova.compute.manager [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.079296] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7fb206e-f890-4fdd-91de-68a82617f952 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.090161] env[63355]: DEBUG nova.compute.manager [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63355) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1022.090161] env[63355]: DEBUG nova.objects.instance [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'flavor' on Instance uuid 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.090545] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1022.090545] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1022.155336] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.163023] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fb6fae-7afc-1a9c-598f-a9422b3e31d5, 'name': SearchDatastore_Task, 'duration_secs': 0.017192} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.163023] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-341e2808-a870-4556-ae8c-920371d97b4b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.172426] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1022.172426] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528eb171-d3bb-8e88-91df-7ebab947bada" [ 1022.172426] env[63355]: _type = "Task" [ 1022.172426] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.181389] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528eb171-d3bb-8e88-91df-7ebab947bada, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.275990] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350033, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.458974] env[63355]: DEBUG oslo_vmware.api [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350035, 'name': ReconfigVM_Task, 'duration_secs': 0.163287} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.459345] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287741', 'volume_id': '64055852-0585-4516-8f95-0996e47ccc04', 'name': 'volume-64055852-0585-4516-8f95-0996e47ccc04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae3961d2-dc5b-4e49-acca-6fb52291f23b', 'attached_at': '', 'detached_at': '', 'volume_id': '64055852-0585-4516-8f95-0996e47ccc04', 'serial': '64055852-0585-4516-8f95-0996e47ccc04'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1022.537160] env[63355]: DEBUG oslo_vmware.api [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350036, 'name': PowerOnVM_Task, 'duration_secs': 0.480816} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.537443] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1022.537682] env[63355]: INFO nova.compute.manager [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Took 8.08 seconds to spawn the instance on the hypervisor. [ 1022.537871] env[63355]: DEBUG nova.compute.manager [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.538961] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb56094c-3a24-4fdb-9f53-4d21b522dd9b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.579693] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.840s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.581937] env[63355]: DEBUG oslo_concurrency.lockutils [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.267s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.584028] env[63355]: DEBUG nova.objects.instance [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lazy-loading 'resources' on Instance uuid cf2d004b-b9dd-4c05-b54d-9509cd1d0b06 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.596494] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.599032] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b1fc906-5ed4-4b41-97ae-6e29e4403d2c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.609198] env[63355]: DEBUG oslo_vmware.api [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1022.609198] env[63355]: value = "task-1350037" [ 1022.609198] env[63355]: _type = "Task" [ 1022.609198] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.620451] env[63355]: DEBUG oslo_vmware.api [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.621504] env[63355]: INFO nova.scheduler.client.report [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Deleted allocations for instance 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7 [ 1022.684454] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528eb171-d3bb-8e88-91df-7ebab947bada, 'name': SearchDatastore_Task, 'duration_secs': 0.011288} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.685518] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.685793] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] fa748748-31ec-431a-a628-5ea179e26fc8/fa748748-31ec-431a-a628-5ea179e26fc8.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1022.686592] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b85c02-1193-4e39-bf17-df28d68d5ac3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.689142] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ea9384e-886a-4b80-9589-3a9ef50d1bb0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.711777] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f51df0-f60c-4d59-9de8-db7a0a4c7f27 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.715495] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1022.715495] env[63355]: value = "task-1350038" [ 1022.715495] env[63355]: _type = "Task" [ 1022.715495] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.722750] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance 'fd55ddfc-f376-4e15-961e-6826f31a7890' progress to 83 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1022.733648] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350038, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.779769] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350033, 'name': CreateSnapshot_Task, 'duration_secs': 1.121201} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.780108] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1022.780931] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720011ac-1f60-4a90-a1ca-b81cf26d7537 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.060791] env[63355]: INFO nova.compute.manager [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Took 21.13 seconds to build instance. [ 1023.131166] env[63355]: DEBUG oslo_vmware.api [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350037, 'name': PowerOffVM_Task, 'duration_secs': 0.233654} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.134807] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a307ad1c-caf2-452d-9af4-13d9b9d076ed tempest-ServerGroupTestJSON-243486216 tempest-ServerGroupTestJSON-243486216-project-member] Lock "9bf49da7-df44-4f26-ac7d-d3a4dab24ce7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.316s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.136563] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1023.136835] env[63355]: DEBUG nova.compute.manager [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.138520] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb5e788-6c23-4815-99f7-270925c8aebc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.230426] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1023.230748] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350038, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.233572] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eccd5b68-6d8c-4cba-a95f-4621ecddb98d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.242234] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1023.242234] env[63355]: value = "task-1350039" [ 1023.242234] env[63355]: _type = "Task" [ 1023.242234] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.253376] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350039, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.300705] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1023.301073] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3182a22c-034e-4572-aaa7-b6fca469b443 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.305917] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a3f006-c13c-4e57-9d1f-709a5ac5c411 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.315612] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39908e60-517e-4847-91a6-994413092c23 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.320187] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1023.320187] env[63355]: value = "task-1350040" [ 1023.320187] env[63355]: _type = "Task" [ 1023.320187] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.351863] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c2d58d-bdb3-4861-a554-a32dcbd4887f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.358630] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350040, 'name': CloneVM_Task} progress is 12%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.364921] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bc826a-565d-4a3c-9924-6e9d1aaed52e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.382399] env[63355]: DEBUG nova.compute.provider_tree [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.514849] env[63355]: DEBUG nova.objects.instance [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'flavor' on Instance uuid ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.563365] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9180b860-bcd8-4193-b112-031cac03987d tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.639s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.653222] env[63355]: DEBUG oslo_concurrency.lockutils [None req-17cba19d-a514-4c05-ba3e-451d288f090c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.575s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.728425] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350038, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.002016} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.728763] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] fa748748-31ec-431a-a628-5ea179e26fc8/fa748748-31ec-431a-a628-5ea179e26fc8.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1023.729022] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.729285] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac21fae9-478a-489f-8676-5597fbc612e7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.738678] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1023.738678] env[63355]: value = "task-1350041" [ 1023.738678] env[63355]: _type = "Task" [ 1023.738678] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.749280] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350041, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.755195] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350039, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.834543] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350040, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.885859] env[63355]: DEBUG nova.scheduler.client.report [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.022915] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6b37359d-90ef-464b-8149-e244a7e3330c tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.340s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.197561] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.198185] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.198185] env[63355]: DEBUG nova.compute.manager [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.199372] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fcd0bf5-31ed-4854-9184-2b707225a1aa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.209974] env[63355]: DEBUG nova.compute.manager [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63355) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1024.210941] env[63355]: DEBUG nova.objects.instance [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'flavor' on Instance uuid ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.258092] env[63355]: DEBUG oslo_vmware.api [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350039, 'name': PowerOnVM_Task, 'duration_secs': 0.935895} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.258410] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350041, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098687} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.258609] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1024.258802] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1179f03f-d635-4561-87a3-a7c578088f0d tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance 'fd55ddfc-f376-4e15-961e-6826f31a7890' progress to 100 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1024.262596] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1024.263687] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f410b92-0719-4acc-b45a-17077deaf1a2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.291840] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] fa748748-31ec-431a-a628-5ea179e26fc8/fa748748-31ec-431a-a628-5ea179e26fc8.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.292219] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f493b83-1647-42c2-8c46-459bf379374a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.315480] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1024.315480] env[63355]: value = "task-1350042" [ 1024.315480] env[63355]: _type = "Task" [ 1024.315480] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.330806] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350042, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.335627] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350040, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.365236] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.365502] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.365683] env[63355]: DEBUG nova.compute.manager [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.367017] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36581b6e-af7d-4b1e-a1c3-edcfc3fe9d4a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.375239] env[63355]: DEBUG nova.compute.manager [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63355) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1024.375860] env[63355]: DEBUG nova.objects.instance [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lazy-loading 'flavor' on Instance uuid 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.391800] env[63355]: DEBUG oslo_concurrency.lockutils [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.394125] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.953s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.394428] env[63355]: DEBUG nova.objects.instance [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lazy-loading 'pci_requests' on Instance uuid 7c153109-b814-4e11-b4f9-7b8cebb853d4 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.416884] env[63355]: INFO nova.scheduler.client.report [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted allocations for instance cf2d004b-b9dd-4c05-b54d-9509cd1d0b06 [ 1024.717049] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.717198] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e3ece59-4328-4a2e-872b-e5daaa512245 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.729541] env[63355]: DEBUG oslo_vmware.api [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1024.729541] env[63355]: value = "task-1350043" [ 1024.729541] env[63355]: _type = "Task" [ 1024.729541] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.738821] env[63355]: DEBUG oslo_vmware.api [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350043, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.831031] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350042, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.838152] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350040, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.883403] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.883628] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ddbf4ba-4e3f-4741-8a18-18f08c7ab49f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.894048] env[63355]: DEBUG oslo_vmware.api [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1024.894048] env[63355]: value = "task-1350044" [ 1024.894048] env[63355]: _type = "Task" [ 1024.894048] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.899860] env[63355]: DEBUG nova.objects.instance [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lazy-loading 'numa_topology' on Instance uuid 7c153109-b814-4e11-b4f9-7b8cebb853d4 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.908772] env[63355]: DEBUG oslo_vmware.api [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350044, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.923850] env[63355]: DEBUG oslo_concurrency.lockutils [None req-95d4d02f-64a4-49bd-90d3-283f9c3fca57 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "cf2d004b-b9dd-4c05-b54d-9509cd1d0b06" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.715s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.119191] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Didn't find any instances for network info cache update. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1025.119437] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1025.119709] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1025.119921] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1025.120134] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1025.120340] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1025.120536] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1025.120693] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1025.120844] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1025.241503] env[63355]: DEBUG oslo_vmware.api [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350043, 'name': PowerOffVM_Task, 'duration_secs': 0.387477} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.242411] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.242608] env[63355]: DEBUG nova.compute.manager [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.243421] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c77c92-7f9f-443d-8fa2-880c97627b34 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.329109] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350042, 'name': ReconfigVM_Task, 'duration_secs': 0.782602} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.329816] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Reconfigured VM instance instance-0000005d to attach disk [datastore2] fa748748-31ec-431a-a628-5ea179e26fc8/fa748748-31ec-431a-a628-5ea179e26fc8.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.330490] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-220d05dc-6aec-4676-93f2-51878e66a83a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.335154] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350040, 'name': CloneVM_Task, 'duration_secs': 1.89099} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.335735] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Created linked-clone VM from snapshot [ 1025.336541] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35ba423-ed98-403d-b691-2811529118d7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.341922] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1025.341922] env[63355]: value = "task-1350045" [ 1025.341922] env[63355]: _type = "Task" [ 1025.341922] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.345854] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Uploading image a0dee740-e030-458a-98f1-353adf5d13fb {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1025.358212] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350045, 'name': Rename_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.360539] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1025.360822] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-acbae687-721c-48b7-bdba-29c066eacce6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.369136] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1025.369136] env[63355]: value = "task-1350046" [ 1025.369136] env[63355]: _type = "Task" [ 1025.369136] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.380844] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350046, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.386418] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "edb14fe7-d444-4fef-8c5d-d5616676e326" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.386673] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "edb14fe7-d444-4fef-8c5d-d5616676e326" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.386891] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "edb14fe7-d444-4fef-8c5d-d5616676e326-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.387094] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "edb14fe7-d444-4fef-8c5d-d5616676e326-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.387271] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "edb14fe7-d444-4fef-8c5d-d5616676e326-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.389666] env[63355]: INFO nova.compute.manager [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Terminating instance [ 1025.392016] env[63355]: DEBUG nova.compute.manager [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1025.392276] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1025.393128] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cbe3621-ceb1-4462-a4f3-0d20beed3b30 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.402996] env[63355]: INFO nova.compute.claims [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.412866] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.413214] env[63355]: DEBUG oslo_vmware.api [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350044, 'name': PowerOffVM_Task, 'duration_secs': 0.255206} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.413699] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4be2ad7-401d-43ca-a481-e3b8718dfecc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.415410] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.415711] env[63355]: DEBUG nova.compute.manager [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.416371] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b23ca6d-e9e1-4162-ad40-e9b815e572e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.428695] env[63355]: DEBUG oslo_vmware.api [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1025.428695] env[63355]: value = "task-1350047" [ 1025.428695] env[63355]: _type = "Task" [ 1025.428695] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.438206] env[63355]: DEBUG oslo_vmware.api [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350047, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.624651] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.722908] env[63355]: DEBUG nova.compute.manager [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Stashing vm_state: stopped {{(pid=63355) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1025.759561] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9a40a968-2608-4ed0-b636-d01af506f502 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.562s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.855418] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350045, 'name': Rename_Task, 'duration_secs': 0.193067} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.855672] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1025.855993] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a2ea2127-02a4-4b35-8a71-5edd92b5cba7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.862724] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1025.862724] env[63355]: value = "task-1350048" [ 1025.862724] env[63355]: _type = "Task" [ 1025.862724] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.871286] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350048, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.879327] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350046, 'name': Destroy_Task} progress is 33%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.935433] env[63355]: DEBUG oslo_concurrency.lockutils [None req-fbb45f83-fd0e-479d-a7a5-e521577c1a83 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.570s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.943340] env[63355]: DEBUG oslo_vmware.api [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350047, 'name': PowerOffVM_Task, 'duration_secs': 0.235658} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.943674] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.944020] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.944295] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1688baa4-39b9-49f6-acd0-59b8083fefbe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.026279] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1026.026714] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1026.027086] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleting the datastore file [datastore2] edb14fe7-d444-4fef-8c5d-d5616676e326 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.027398] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10b2dee8-4825-4669-b7e2-0f12d7dc40fc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.035798] env[63355]: DEBUG oslo_vmware.api [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1026.035798] env[63355]: value = "task-1350050" [ 1026.035798] env[63355]: _type = "Task" [ 1026.035798] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.045824] env[63355]: DEBUG oslo_vmware.api [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350050, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.245550] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.373638] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350048, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.385661] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350046, 'name': Destroy_Task, 'duration_secs': 0.549719} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.386378] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Destroyed the VM [ 1026.386378] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1026.386750] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1ce81f80-95bd-41ba-944d-e0c6e7fb1145 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.395293] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1026.395293] env[63355]: value = "task-1350051" [ 1026.395293] env[63355]: _type = "Task" [ 1026.395293] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.404773] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350051, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.546545] env[63355]: DEBUG oslo_vmware.api [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350050, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205225} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.549213] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.549412] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1026.549597] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1026.549776] env[63355]: INFO nova.compute.manager [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1026.550034] env[63355]: DEBUG oslo.service.loopingcall [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.554439] env[63355]: DEBUG nova.compute.manager [-] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.554561] env[63355]: DEBUG nova.network.neutron [-] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1026.567585] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "fd55ddfc-f376-4e15-961e-6826f31a7890" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.567992] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.568239] env[63355]: DEBUG nova.compute.manager [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Going to confirm migration 1 {{(pid=63355) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1026.584325] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2d24c6-6965-47ec-aff4-74b0ac70027b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.592042] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba567352-0b92-4451-9d4a-6a6be00263dd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.622704] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f117f8-58ac-4106-a0fe-191b6da897f9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.632247] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6d51dc-a980-462a-ab2f-06271768d0a3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.650326] env[63355]: DEBUG nova.compute.provider_tree [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.770020] env[63355]: DEBUG nova.objects.instance [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'flavor' on Instance uuid ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.872960] env[63355]: DEBUG oslo_vmware.api [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350048, 'name': PowerOnVM_Task, 'duration_secs': 0.530361} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.873277] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1026.873490] env[63355]: INFO nova.compute.manager [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Took 10.08 seconds to spawn the instance on the hypervisor. [ 1026.873711] env[63355]: DEBUG nova.compute.manager [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1026.874638] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0b89d1-df3a-4387-a40e-5aa5d4e535e0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.905260] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350051, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.986595] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.986903] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.987152] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.987341] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.987539] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.989904] env[63355]: INFO nova.compute.manager [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Terminating instance [ 1026.991685] env[63355]: DEBUG nova.compute.manager [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1026.991878] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1026.992681] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715807e0-b408-47b8-a31d-f0ed01102cf1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.998432] env[63355]: DEBUG nova.compute.manager [req-c7f60350-cc2a-4664-8484-90878dc80392 req-2f894e66-9f39-4c6b-9153-e3a99cebf0c1 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Received event network-vif-deleted-677de229-5411-464b-9680-0080e5794772 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.998529] env[63355]: INFO nova.compute.manager [req-c7f60350-cc2a-4664-8484-90878dc80392 req-2f894e66-9f39-4c6b-9153-e3a99cebf0c1 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Neutron deleted interface 677de229-5411-464b-9680-0080e5794772; detaching it from the instance and deleting it from the info cache [ 1026.998803] env[63355]: DEBUG nova.network.neutron [req-c7f60350-cc2a-4664-8484-90878dc80392 req-2f894e66-9f39-4c6b-9153-e3a99cebf0c1 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.002252] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1027.002653] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66e79892-7304-4142-93d0-774095e01c4f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.074873] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1027.075112] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1027.075305] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleting the datastore file [datastore2] 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.077859] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea9d2fac-c620-4ada-acf2-8dc10021ade6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.086909] env[63355]: DEBUG oslo_vmware.api [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1027.086909] env[63355]: value = "task-1350053" [ 1027.086909] env[63355]: _type = "Task" [ 1027.086909] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.100126] env[63355]: DEBUG oslo_vmware.api [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350053, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.153344] env[63355]: DEBUG nova.scheduler.client.report [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.182628] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.182825] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.183016] env[63355]: DEBUG nova.network.neutron [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.183221] env[63355]: DEBUG nova.objects.instance [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lazy-loading 'info_cache' on Instance uuid fd55ddfc-f376-4e15-961e-6826f31a7890 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.276693] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.276693] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquired lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.276693] env[63355]: DEBUG nova.network.neutron [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.276693] env[63355]: DEBUG nova.objects.instance [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'info_cache' on Instance uuid ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.393753] env[63355]: INFO nova.compute.manager [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Took 24.23 seconds to build instance. [ 1027.405761] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350051, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.471662] env[63355]: DEBUG nova.network.neutron [-] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.503535] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24488aa1-cbe7-45a6-97a5-9cbe04b5a1e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.513640] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73e8296-6072-409a-874e-1ba6f09ba95b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.542250] env[63355]: DEBUG nova.compute.manager [req-c7f60350-cc2a-4664-8484-90878dc80392 req-2f894e66-9f39-4c6b-9153-e3a99cebf0c1 service nova] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Detach interface failed, port_id=677de229-5411-464b-9680-0080e5794772, reason: Instance edb14fe7-d444-4fef-8c5d-d5616676e326 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1027.598640] env[63355]: DEBUG oslo_vmware.api [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350053, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.263846} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.598969] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.599247] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1027.599535] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1027.599816] env[63355]: INFO nova.compute.manager [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1027.600246] env[63355]: DEBUG oslo.service.loopingcall [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.600529] env[63355]: DEBUG nova.compute.manager [-] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1027.600669] env[63355]: DEBUG nova.network.neutron [-] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1027.658917] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.265s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.660871] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.037s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.661057] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.661231] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1027.661871] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.417s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.663747] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e21516-5ecf-4759-893a-efe994d34dac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.672619] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415b7c14-09ca-41d4-a4f9-558a6a086f0f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.690429] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd1da5c-0149-4819-9fff-6841cced39b8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.694992] env[63355]: INFO nova.network.neutron [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updating port 43ebe841-d4c8-446a-981c-519bef977228 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1027.700192] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63212cd8-7221-45c9-a1cd-63b4baa9e053 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.732350] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180038MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1027.732552] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.779510] env[63355]: DEBUG nova.objects.base [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1027.895640] env[63355]: DEBUG oslo_concurrency.lockutils [None req-023c67c5-a0dd-4c29-ab05-87be408a6be0 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "fa748748-31ec-431a-a628-5ea179e26fc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.741s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.906421] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350051, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.974746] env[63355]: INFO nova.compute.manager [-] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Took 1.42 seconds to deallocate network for instance. [ 1028.170150] env[63355]: INFO nova.compute.claims [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.407508] env[63355]: DEBUG oslo_vmware.api [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350051, 'name': RemoveSnapshot_Task, 'duration_secs': 1.588139} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.407851] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1028.481206] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.489536] env[63355]: DEBUG nova.network.neutron [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updating instance_info_cache with network_info: [{"id": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "address": "fa:16:3e:fb:af:dc", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4c4170a-d2", "ovs_interfaceid": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.631865] env[63355]: DEBUG nova.network.neutron [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance_info_cache with network_info: [{"id": "52d11eb4-329c-4989-9498-377fd46350cb", "address": "fa:16:3e:64:c5:93", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52d11eb4-32", "ovs_interfaceid": "52d11eb4-329c-4989-9498-377fd46350cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.638050] env[63355]: DEBUG nova.network.neutron [-] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.677792] env[63355]: INFO nova.compute.resource_tracker [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating resource usage from migration 0d5cde65-ecd9-4f5b-b4bf-dfd872c38a3f [ 1028.830792] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2600560f-594f-4e16-9612-5cacb3c96568 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.839362] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f63ed8-6aaa-486c-985c-df54180636e3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.869696] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2814a417-8785-4282-a747-2c2fb57b044b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.878095] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee7f979-bfdb-4371-9f92-6f0b6d6c8e02 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.893785] env[63355]: DEBUG nova.compute.provider_tree [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.913534] env[63355]: WARNING nova.compute.manager [None req-544eef02-9ee6-4831-b56c-d40d34cead11 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Image not found during snapshot: nova.exception.ImageNotFound: Image a0dee740-e030-458a-98f1-353adf5d13fb could not be found. [ 1028.993066] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Releasing lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.056637] env[63355]: DEBUG nova.compute.manager [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Received event network-vif-deleted-6d3a3c43-ae33-47e1-9e1b-bf0d9750e565 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.056885] env[63355]: DEBUG nova.compute.manager [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received event network-changed-aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.057524] env[63355]: DEBUG nova.compute.manager [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing instance network info cache due to event network-changed-aa5a40b6-5d12-4d62-88d2-32a26a65a241. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1029.057935] env[63355]: DEBUG oslo_concurrency.lockutils [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] Acquiring lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.058216] env[63355]: DEBUG oslo_concurrency.lockutils [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] Acquired lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.058399] env[63355]: DEBUG nova.network.neutron [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing network info cache for port aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1029.134517] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-fd55ddfc-f376-4e15-961e-6826f31a7890" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.134806] env[63355]: DEBUG nova.objects.instance [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lazy-loading 'migration_context' on Instance uuid fd55ddfc-f376-4e15-961e-6826f31a7890 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.139988] env[63355]: INFO nova.compute.manager [-] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Took 1.54 seconds to deallocate network for instance. [ 1029.147149] env[63355]: DEBUG nova.compute.manager [req-cd2567a9-7f4c-42e8-82f1-b3304aab12e0 req-94733545-d25f-4f03-a720-a40696ff54a6 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received event network-vif-plugged-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.147363] env[63355]: DEBUG oslo_concurrency.lockutils [req-cd2567a9-7f4c-42e8-82f1-b3304aab12e0 req-94733545-d25f-4f03-a720-a40696ff54a6 service nova] Acquiring lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.147616] env[63355]: DEBUG oslo_concurrency.lockutils [req-cd2567a9-7f4c-42e8-82f1-b3304aab12e0 req-94733545-d25f-4f03-a720-a40696ff54a6 service nova] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.147801] env[63355]: DEBUG oslo_concurrency.lockutils [req-cd2567a9-7f4c-42e8-82f1-b3304aab12e0 req-94733545-d25f-4f03-a720-a40696ff54a6 service nova] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.147972] env[63355]: DEBUG nova.compute.manager [req-cd2567a9-7f4c-42e8-82f1-b3304aab12e0 req-94733545-d25f-4f03-a720-a40696ff54a6 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] No waiting events found dispatching network-vif-plugged-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1029.148189] env[63355]: WARNING nova.compute.manager [req-cd2567a9-7f4c-42e8-82f1-b3304aab12e0 req-94733545-d25f-4f03-a720-a40696ff54a6 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received unexpected event network-vif-plugged-43ebe841-d4c8-446a-981c-519bef977228 for instance with vm_state shelved_offloaded and task_state spawning. [ 1029.238461] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.238461] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.238602] env[63355]: DEBUG nova.network.neutron [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1029.397354] env[63355]: DEBUG nova.scheduler.client.report [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.497191] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1029.497586] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-640d3a46-cb39-498f-8cf3-1bae5802a1f4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.506508] env[63355]: DEBUG oslo_vmware.api [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1029.506508] env[63355]: value = "task-1350054" [ 1029.506508] env[63355]: _type = "Task" [ 1029.506508] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.515253] env[63355]: DEBUG oslo_vmware.api [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350054, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.643805] env[63355]: DEBUG nova.objects.base [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1029.643805] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b8e58f-861f-4e9c-b255-dd878c7d1917 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.650155] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.669723] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9590b3f4-d897-48cc-b3cf-ec962333067b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.677192] env[63355]: DEBUG oslo_vmware.api [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1029.677192] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a85084-f908-09f7-87e8-b75d4972c190" [ 1029.677192] env[63355]: _type = "Task" [ 1029.677192] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.687275] env[63355]: DEBUG oslo_vmware.api [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a85084-f908-09f7-87e8-b75d4972c190, 'name': SearchDatastore_Task, 'duration_secs': 0.007384} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.687767] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.776046] env[63355]: DEBUG nova.network.neutron [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updated VIF entry in instance network info cache for port aa5a40b6-5d12-4d62-88d2-32a26a65a241. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1029.776046] env[63355]: DEBUG nova.network.neutron [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [{"id": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "address": "fa:16:3e:08:9e:2a", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5a40b6-5d", "ovs_interfaceid": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.906852] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.243s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.906852] env[63355]: INFO nova.compute.manager [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Migrating [ 1029.913469] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.181s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.919271] env[63355]: DEBUG oslo_concurrency.lockutils [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.919271] env[63355]: DEBUG oslo_concurrency.lockutils [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.919791] env[63355]: DEBUG oslo_concurrency.lockutils [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.920165] env[63355]: DEBUG oslo_concurrency.lockutils [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.920477] env[63355]: DEBUG oslo_concurrency.lockutils [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.927019] env[63355]: INFO nova.compute.manager [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Terminating instance [ 1029.934743] env[63355]: DEBUG nova.compute.manager [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1029.936442] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1029.938655] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2107d35e-6f68-49eb-8174-40540d6cb7f7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.948837] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.949311] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0100a13-24d9-427e-9ba1-fcf49f4de0e9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.958536] env[63355]: DEBUG oslo_vmware.api [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1029.958536] env[63355]: value = "task-1350055" [ 1029.958536] env[63355]: _type = "Task" [ 1029.958536] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.967913] env[63355]: DEBUG oslo_vmware.api [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350055, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.016917] env[63355]: DEBUG oslo_vmware.api [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350054, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.281859] env[63355]: DEBUG oslo_concurrency.lockutils [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] Releasing lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.282159] env[63355]: DEBUG nova.compute.manager [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Received event network-changed-546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.282337] env[63355]: DEBUG nova.compute.manager [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Refreshing instance network info cache due to event network-changed-546aac3d-2d71-48bd-84ac-6161e42a90dd. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1030.282559] env[63355]: DEBUG oslo_concurrency.lockutils [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] Acquiring lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.282707] env[63355]: DEBUG oslo_concurrency.lockutils [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] Acquired lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.282871] env[63355]: DEBUG nova.network.neutron [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Refreshing network info cache for port 546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1030.358281] env[63355]: DEBUG nova.network.neutron [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updating instance_info_cache with network_info: [{"id": "43ebe841-d4c8-446a-981c-519bef977228", "address": "fa:16:3e:97:62:7e", "network": {"id": "e19b4619-6e81-4205-838d-9ee7e7ae44d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1968447346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "744ce7dc0ab4463fa461cbac87e51904", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43ebe841-d4", "ovs_interfaceid": "43ebe841-d4c8-446a-981c-519bef977228", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.440084] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.440298] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.440917] env[63355]: DEBUG nova.network.neutron [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1030.468904] env[63355]: DEBUG oslo_vmware.api [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350055, 'name': PowerOffVM_Task, 'duration_secs': 0.229674} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.469210] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1030.469394] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1030.469701] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82491aa4-a4e9-4a6c-8889-465459cad470 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.516331] env[63355]: DEBUG oslo_vmware.api [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350054, 'name': PowerOnVM_Task, 'duration_secs': 0.837261} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.516741] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1030.516822] env[63355]: DEBUG nova.compute.manager [None req-e3e7d6b8-74e9-4566-959b-a875cbadc8dd tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.517590] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a60e46-6ba0-43ca-b19f-0f95bab5a852 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.543268] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.543268] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.543268] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleting the datastore file [datastore1] 40c48196-6543-4c96-9b8c-3b3c9fbccfe4 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.543542] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94183ba1-0316-48e9-a6b3-9f7b85af406d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.553328] env[63355]: DEBUG oslo_vmware.api [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for the task: (returnval){ [ 1030.553328] env[63355]: value = "task-1350057" [ 1030.553328] env[63355]: _type = "Task" [ 1030.553328] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.560786] env[63355]: DEBUG oslo_vmware.api [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350057, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.863506] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Releasing lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.889420] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='ae2058312cedac42f9c4bc593495fccf',container_format='bare',created_at=2024-10-24T02:31:27Z,direct_url=,disk_format='vmdk',id=c58460ed-ef70-483a-87f9-6152e2a0d383,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-788921884-shelved',owner='744ce7dc0ab4463fa461cbac87e51904',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-24T02:31:43Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.889678] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.889841] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.890038] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.890197] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.890350] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.890560] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.890723] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.890891] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.891070] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.891247] env[63355]: DEBUG nova.virt.hardware [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.892080] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa54a1a5-902b-4a4b-a92b-e470a17b4df8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.900321] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8bcc7f-c3ad-414f-b1d9-6f88d8949620 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.914034] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:62:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16f09e8c-5240-4839-80cc-62ec29700bd2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43ebe841-d4c8-446a-981c-519bef977228', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.920753] env[63355]: DEBUG oslo.service.loopingcall [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.923220] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.923328] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-725d8b5f-50f1-408e-ac5a-1a0f735362cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.943173] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Applying migration context for instance 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 as it has an incoming, in-progress migration 0d5cde65-ecd9-4f5b-b4bf-dfd872c38a3f. Migration status is pre-migrating {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1030.943470] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Applying migration context for instance fd55ddfc-f376-4e15-961e-6826f31a7890 as it has an incoming, in-progress migration cab18143-dc8b-4891-851e-80d74061526f. Migration status is confirming {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1030.945044] env[63355]: INFO nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating resource usage from migration 0d5cde65-ecd9-4f5b-b4bf-dfd872c38a3f [ 1030.945044] env[63355]: INFO nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating resource usage from migration cab18143-dc8b-4891-851e-80d74061526f [ 1030.949645] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.949645] env[63355]: value = "task-1350058" [ 1030.949645] env[63355]: _type = "Task" [ 1030.949645] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.961515] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350058, 'name': CreateVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.971228] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance ae3961d2-dc5b-4e49-acca-6fb52291f23b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1030.971336] env[63355]: WARNING nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance edb14fe7-d444-4fef-8c5d-d5616676e326 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1030.971454] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance da3cb83c-6368-49c0-9b11-0498221e3c0f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1030.971575] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 40c48196-6543-4c96-9b8c-3b3c9fbccfe4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1030.971703] env[63355]: WARNING nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1030.971826] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Migration cab18143-dc8b-4891-851e-80d74061526f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1030.971943] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance fd55ddfc-f376-4e15-961e-6826f31a7890 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1030.972074] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance fa748748-31ec-431a-a628-5ea179e26fc8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1030.972193] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7c153109-b814-4e11-b4f9-7b8cebb853d4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1030.972307] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Migration 0d5cde65-ecd9-4f5b-b4bf-dfd872c38a3f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1030.972493] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1030.972614] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1030.973029] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2368MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1031.073632] env[63355]: DEBUG oslo_vmware.api [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Task: {'id': task-1350057, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149835} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.073632] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1031.073953] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1031.074074] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1031.074392] env[63355]: INFO nova.compute.manager [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1031.074855] env[63355]: DEBUG oslo.service.loopingcall [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.079144] env[63355]: DEBUG nova.compute.manager [-] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1031.079270] env[63355]: DEBUG nova.network.neutron [-] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1031.105080] env[63355]: DEBUG nova.network.neutron [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updated VIF entry in instance network info cache for port 546aac3d-2d71-48bd-84ac-6161e42a90dd. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.105080] env[63355]: DEBUG nova.network.neutron [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updating instance_info_cache with network_info: [{"id": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "address": "fa:16:3e:40:51:98", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546aac3d-2d", "ovs_interfaceid": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.127064] env[63355]: DEBUG nova.compute.manager [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Received event network-changed-546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.127064] env[63355]: DEBUG nova.compute.manager [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Refreshing instance network info cache due to event network-changed-546aac3d-2d71-48bd-84ac-6161e42a90dd. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1031.128322] env[63355]: DEBUG oslo_concurrency.lockutils [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] Acquiring lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.168737] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820cedcc-cffd-4d2e-8d82-19f9946553a5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.179673] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17396a57-8d04-4665-b0d3-ca6146305b84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.187842] env[63355]: DEBUG nova.compute.manager [req-96b29d76-d591-4428-a142-a657c24e168e req-1c56478a-96a8-4ade-bc7f-fd55814980c4 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received event network-changed-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.187842] env[63355]: DEBUG nova.compute.manager [req-96b29d76-d591-4428-a142-a657c24e168e req-1c56478a-96a8-4ade-bc7f-fd55814980c4 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Refreshing instance network info cache due to event network-changed-43ebe841-d4c8-446a-981c-519bef977228. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1031.187842] env[63355]: DEBUG oslo_concurrency.lockutils [req-96b29d76-d591-4428-a142-a657c24e168e req-1c56478a-96a8-4ade-bc7f-fd55814980c4 service nova] Acquiring lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.187842] env[63355]: DEBUG oslo_concurrency.lockutils [req-96b29d76-d591-4428-a142-a657c24e168e req-1c56478a-96a8-4ade-bc7f-fd55814980c4 service nova] Acquired lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.187842] env[63355]: DEBUG nova.network.neutron [req-96b29d76-d591-4428-a142-a657c24e168e req-1c56478a-96a8-4ade-bc7f-fd55814980c4 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Refreshing network info cache for port 43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1031.224788] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f1dc63-6e7a-4992-bb9b-533cdeead45c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.234566] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dbd13af-b119-40b0-935a-2ba6ebbb0913 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.250923] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.298728] env[63355]: DEBUG nova.network.neutron [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance_info_cache with network_info: [{"id": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "address": "fa:16:3e:b3:26:57", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap976ab0e0-5c", "ovs_interfaceid": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.461768] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350058, 'name': CreateVM_Task, 'duration_secs': 0.415554} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.462535] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.463264] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.463436] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.463856] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1031.464383] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4afa394-6237-4e15-80cd-7168fef65d85 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.469501] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1031.469501] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524298ff-04ae-48b1-aa5a-3fe7c8bb7e7c" [ 1031.469501] env[63355]: _type = "Task" [ 1031.469501] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.487243] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.487555] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Processing image c58460ed-ef70-483a-87f9-6152e2a0d383 {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.487812] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383/c58460ed-ef70-483a-87f9-6152e2a0d383.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.487969] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383/c58460ed-ef70-483a-87f9-6152e2a0d383.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.488169] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.488431] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b88985fd-f930-421b-ba02-2b31be327d9f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.506648] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.506873] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.507637] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1990eab6-7dbf-409f-bc48-a33472f72773 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.513127] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1031.513127] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f9a148-7c32-48e7-e8a6-ad033dd7742f" [ 1031.513127] env[63355]: _type = "Task" [ 1031.513127] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.523471] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f9a148-7c32-48e7-e8a6-ad033dd7742f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.550090] env[63355]: DEBUG nova.network.neutron [req-96b29d76-d591-4428-a142-a657c24e168e req-1c56478a-96a8-4ade-bc7f-fd55814980c4 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updated VIF entry in instance network info cache for port 43ebe841-d4c8-446a-981c-519bef977228. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.550469] env[63355]: DEBUG nova.network.neutron [req-96b29d76-d591-4428-a142-a657c24e168e req-1c56478a-96a8-4ade-bc7f-fd55814980c4 service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updating instance_info_cache with network_info: [{"id": "43ebe841-d4c8-446a-981c-519bef977228", "address": "fa:16:3e:97:62:7e", "network": {"id": "e19b4619-6e81-4205-838d-9ee7e7ae44d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1968447346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "744ce7dc0ab4463fa461cbac87e51904", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43ebe841-d4", "ovs_interfaceid": "43ebe841-d4c8-446a-981c-519bef977228", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.607656] env[63355]: DEBUG oslo_concurrency.lockutils [req-8c72d299-b247-4ccd-bd81-59bb4dc01c65 req-debf5fa5-29f6-4a7e-bf04-20d8024edd66 service nova] Releasing lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.608112] env[63355]: DEBUG oslo_concurrency.lockutils [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] Acquired lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.608318] env[63355]: DEBUG nova.network.neutron [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Refreshing network info cache for port 546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1031.755041] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.801986] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.956903] env[63355]: DEBUG nova.network.neutron [-] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.022938] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Preparing fetch location {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1032.023219] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Fetch image to [datastore2] OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac/OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac.vmdk {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1032.023409] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Downloading stream optimized image c58460ed-ef70-483a-87f9-6152e2a0d383 to [datastore2] OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac/OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac.vmdk on the data store datastore2 as vApp {{(pid=63355) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1032.023585] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Downloading image file data c58460ed-ef70-483a-87f9-6152e2a0d383 to the ESX as VM named 'OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac' {{(pid=63355) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1032.068622] env[63355]: DEBUG oslo_concurrency.lockutils [req-96b29d76-d591-4428-a142-a657c24e168e req-1c56478a-96a8-4ade-bc7f-fd55814980c4 service nova] Releasing lock "refresh_cache-7c153109-b814-4e11-b4f9-7b8cebb853d4" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.103144] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1032.103144] env[63355]: value = "resgroup-9" [ 1032.103144] env[63355]: _type = "ResourcePool" [ 1032.103144] env[63355]: }. {{(pid=63355) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1032.103392] env[63355]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-72c267b3-b229-4a09-8054-b92fa61601ef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.127738] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lease: (returnval){ [ 1032.127738] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fbfe42-0568-3462-eeff-8e50078094d7" [ 1032.127738] env[63355]: _type = "HttpNfcLease" [ 1032.127738] env[63355]: } obtained for vApp import into resource pool (val){ [ 1032.127738] env[63355]: value = "resgroup-9" [ 1032.127738] env[63355]: _type = "ResourcePool" [ 1032.127738] env[63355]: }. {{(pid=63355) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1032.128129] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the lease: (returnval){ [ 1032.128129] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fbfe42-0568-3462-eeff-8e50078094d7" [ 1032.128129] env[63355]: _type = "HttpNfcLease" [ 1032.128129] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1032.135289] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1032.135289] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fbfe42-0568-3462-eeff-8e50078094d7" [ 1032.135289] env[63355]: _type = "HttpNfcLease" [ 1032.135289] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1032.261603] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1032.261603] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.348s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.261603] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.780s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.261603] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.263886] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.614s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.264269] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.266259] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.579s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.302578] env[63355]: INFO nova.scheduler.client.report [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted allocations for instance 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07 [ 1032.312605] env[63355]: INFO nova.scheduler.client.report [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted allocations for instance edb14fe7-d444-4fef-8c5d-d5616676e326 [ 1032.440794] env[63355]: DEBUG nova.network.neutron [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updated VIF entry in instance network info cache for port 546aac3d-2d71-48bd-84ac-6161e42a90dd. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1032.441248] env[63355]: DEBUG nova.network.neutron [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updating instance_info_cache with network_info: [{"id": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "address": "fa:16:3e:40:51:98", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546aac3d-2d", "ovs_interfaceid": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.459695] env[63355]: INFO nova.compute.manager [-] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Took 1.38 seconds to deallocate network for instance. [ 1032.636668] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1032.636668] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fbfe42-0568-3462-eeff-8e50078094d7" [ 1032.636668] env[63355]: _type = "HttpNfcLease" [ 1032.636668] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1032.822978] env[63355]: DEBUG oslo_concurrency.lockutils [None req-54ef36c6-6d17-4622-9cb9-1fe50c012814 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "2214ffc8-d0fd-49f3-91c1-74c13ef7bc07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.836s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.827290] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3b81696e-a469-407b-9f4e-41247ebf2aa4 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "edb14fe7-d444-4fef-8c5d-d5616676e326" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.440s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.902590] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4786c180-866a-494a-a481-5e1182d217b7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.912250] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c724b4b-cf66-4589-ab42-83e0180e1599 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.943531] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5f12f8-c185-40e7-bf78-c979c1840db8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.946337] env[63355]: DEBUG oslo_concurrency.lockutils [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] Releasing lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.946628] env[63355]: DEBUG nova.compute.manager [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received event network-changed-aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.946830] env[63355]: DEBUG nova.compute.manager [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing instance network info cache due to event network-changed-aa5a40b6-5d12-4d62-88d2-32a26a65a241. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1032.947122] env[63355]: DEBUG oslo_concurrency.lockutils [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] Acquiring lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.947245] env[63355]: DEBUG oslo_concurrency.lockutils [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] Acquired lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.947360] env[63355]: DEBUG nova.network.neutron [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing network info cache for port aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1032.954164] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8257e909-5813-4c9f-9f4d-3e85568a9ecc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.969145] env[63355]: DEBUG oslo_concurrency.lockutils [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.969612] env[63355]: DEBUG nova.compute.provider_tree [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.137986] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1033.137986] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fbfe42-0568-3462-eeff-8e50078094d7" [ 1033.137986] env[63355]: _type = "HttpNfcLease" [ 1033.137986] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1033.212262] env[63355]: DEBUG nova.compute.manager [req-455361ce-97d1-4802-b8ad-c166090b94a6 req-c0b31155-f6e8-400b-b2f9-be634b137823 service nova] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Received event network-vif-deleted-a0dfd814-02e8-40d1-acbc-d9447b89a9d6 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.332211] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d6f36f-3b08-4bd1-8a95-4584397e1f8e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.351996] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance '537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5' progress to 0 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1033.472009] env[63355]: DEBUG nova.scheduler.client.report [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.642566] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1033.642566] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fbfe42-0568-3462-eeff-8e50078094d7" [ 1033.642566] env[63355]: _type = "HttpNfcLease" [ 1033.642566] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1033.642933] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1033.642933] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fbfe42-0568-3462-eeff-8e50078094d7" [ 1033.642933] env[63355]: _type = "HttpNfcLease" [ 1033.642933] env[63355]: }. {{(pid=63355) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1033.643709] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b35e9cb-0adb-40ae-99db-5e74d6f39f69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.654843] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a1485-e83a-70b5-23fa-32981da11264/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1033.655052] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a1485-e83a-70b5-23fa-32981da11264/disk-0.vmdk. {{(pid=63355) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1033.718145] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "e271159e-14ed-4a88-a7b8-23d404d77985" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.718414] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "e271159e-14ed-4a88-a7b8-23d404d77985" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.725646] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9f6ed97b-e455-4e5c-9925-b9709488c9ed {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.826440] env[63355]: DEBUG nova.network.neutron [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updated VIF entry in instance network info cache for port aa5a40b6-5d12-4d62-88d2-32a26a65a241. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1033.827020] env[63355]: DEBUG nova.network.neutron [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [{"id": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "address": "fa:16:3e:08:9e:2a", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5a40b6-5d", "ovs_interfaceid": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.858029] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.858544] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae8f999b-8880-4214-a6e7-0b02ee53f845 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.867967] env[63355]: DEBUG oslo_vmware.api [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1033.867967] env[63355]: value = "task-1350060" [ 1033.867967] env[63355]: _type = "Task" [ 1033.867967] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.880973] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] VM already powered off {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1033.881230] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance '537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5' progress to 17 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1034.225785] env[63355]: DEBUG nova.compute.manager [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1034.330374] env[63355]: DEBUG oslo_concurrency.lockutils [req-811aff3f-4a9e-44ca-9a08-7b9e8faab264 req-d283ba60-8dc1-493b-b12f-133e6bd40bc8 service nova] Releasing lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.388208] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1034.388471] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1034.388635] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1034.388824] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1034.388995] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1034.389588] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1034.389588] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1034.389729] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1034.389783] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1034.389951] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1034.390146] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1034.395705] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68422eeb-13a3-463a-a049-f4aa060ae4c5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.417778] env[63355]: DEBUG oslo_vmware.api [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1034.417778] env[63355]: value = "task-1350061" [ 1034.417778] env[63355]: _type = "Task" [ 1034.417778] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.429599] env[63355]: DEBUG oslo_vmware.api [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350061, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.453063] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "e398ee67-eb02-4256-9120-06111f110692" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.453390] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "e398ee67-eb02-4256-9120-06111f110692" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.483366] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.217s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.486292] env[63355]: DEBUG oslo_concurrency.lockutils [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.517s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.486572] env[63355]: DEBUG nova.objects.instance [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lazy-loading 'resources' on Instance uuid 40c48196-6543-4c96-9b8c-3b3c9fbccfe4 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.749762] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.894922] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Completed reading data from the image iterator. {{(pid=63355) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1034.895228] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a1485-e83a-70b5-23fa-32981da11264/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1034.896276] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d11200-094e-4ec0-bce5-c19aaa63fc3a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.903829] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a1485-e83a-70b5-23fa-32981da11264/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1034.904070] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a1485-e83a-70b5-23fa-32981da11264/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1034.904350] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-4ad88da6-4be3-45f5-ae2a-753d43eac3df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.930362] env[63355]: DEBUG oslo_vmware.api [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350061, 'name': ReconfigVM_Task, 'duration_secs': 0.154036} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.930747] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance '537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5' progress to 33 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1034.957328] env[63355]: DEBUG nova.compute.manager [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1035.045511] env[63355]: INFO nova.scheduler.client.report [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted allocation for migration cab18143-dc8b-4891-851e-80d74061526f [ 1035.096841] env[63355]: DEBUG oslo_vmware.rw_handles [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a1485-e83a-70b5-23fa-32981da11264/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1035.097220] env[63355]: INFO nova.virt.vmwareapi.images [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Downloaded image file data c58460ed-ef70-483a-87f9-6152e2a0d383 [ 1035.098688] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773a4929-45fb-4d25-acd5-8aa6ad4bdb8d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.117215] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a57fe95-1a2f-426d-bf8f-7667e357f688 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.143877] env[63355]: INFO nova.virt.vmwareapi.images [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] The imported VM was unregistered [ 1035.146727] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Caching image {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1035.146977] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Creating directory with path [datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383 {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.147289] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1010299-f3d5-4521-81ef-23f86eddaa31 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.150707] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3bc256-eb5d-4053-a165-01fb0d68bcb7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.159980] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e445f27-3a5a-4c01-910b-9f43a398ae4f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.191889] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc01de4-87ca-4998-95a3-17ea766e7a61 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.201307] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ffa1c17-2943-4583-a780-4691325d72f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.215142] env[63355]: DEBUG nova.compute.provider_tree [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.320952] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Created directory with path [datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383 {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.321150] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac/OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac.vmdk to [datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383/c58460ed-ef70-483a-87f9-6152e2a0d383.vmdk. {{(pid=63355) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1035.321397] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-881fd5ff-5209-47f0-97d2-b1be958cac77 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.330354] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1035.330354] env[63355]: value = "task-1350063" [ 1035.330354] env[63355]: _type = "Task" [ 1035.330354] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.338658] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350063, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.440627] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1035.441030] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1035.441205] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1035.441396] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1035.441549] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1035.441745] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1035.441968] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1035.442155] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1035.442329] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1035.442500] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1035.442680] env[63355]: DEBUG nova.virt.hardware [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1035.448070] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Reconfiguring VM instance instance-00000046 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1035.448371] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1efe6096-b7a5-42be-9b58-2dc38e4031ba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.493550] env[63355]: DEBUG oslo_vmware.api [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1035.493550] env[63355]: value = "task-1350064" [ 1035.493550] env[63355]: _type = "Task" [ 1035.493550] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.493550] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.493550] env[63355]: DEBUG oslo_vmware.api [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350064, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.551338] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8bb5a2a2-adaa-4186-9f0b-8995dc0f27ec tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.983s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.719094] env[63355]: DEBUG nova.scheduler.client.report [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.841485] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350063, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.982951] env[63355]: DEBUG oslo_vmware.api [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350064, 'name': ReconfigVM_Task, 'duration_secs': 0.170054} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.983325] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Reconfigured VM instance instance-00000046 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1035.984046] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2962a14a-def8-492b-a5c1-1b66d03cbb74 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.006655] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5/537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.007102] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4864a07e-6d52-475c-a375-8562c78691e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.027180] env[63355]: DEBUG oslo_vmware.api [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1036.027180] env[63355]: value = "task-1350065" [ 1036.027180] env[63355]: _type = "Task" [ 1036.027180] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.040807] env[63355]: DEBUG oslo_vmware.api [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350065, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.224124] env[63355]: DEBUG oslo_concurrency.lockutils [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.738s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.226955] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.477s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.229357] env[63355]: INFO nova.compute.claims [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.244055] env[63355]: INFO nova.scheduler.client.report [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Deleted allocations for instance 40c48196-6543-4c96-9b8c-3b3c9fbccfe4 [ 1036.343120] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350063, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.541261] env[63355]: DEBUG oslo_vmware.api [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350065, 'name': ReconfigVM_Task, 'duration_secs': 0.313232} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.541566] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5/537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.541840] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance '537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5' progress to 50 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1036.757203] env[63355]: DEBUG oslo_concurrency.lockutils [None req-090255ff-6091-4d12-a478-1d46b21706b7 tempest-ImagesTestJSON-2064562919 tempest-ImagesTestJSON-2064562919-project-member] Lock "40c48196-6543-4c96-9b8c-3b3c9fbccfe4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.838s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.846407] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350063, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.051019] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815e2c19-6c12-4278-833b-4c1b4c11562a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.076356] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "interface-da3cb83c-6368-49c0-9b11-0498221e3c0f-563020c5-4b41-4820-9807-82cf864fcade" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.076356] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-da3cb83c-6368-49c0-9b11-0498221e3c0f-563020c5-4b41-4820-9807-82cf864fcade" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.076356] env[63355]: DEBUG nova.objects.instance [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'flavor' on Instance uuid da3cb83c-6368-49c0-9b11-0498221e3c0f {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.078445] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8623d3e4-5bfb-4a6b-844e-1d27a16643b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.100715] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance '537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5' progress to 67 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1037.303563] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "fd55ddfc-f376-4e15-961e-6826f31a7890" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.303957] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.304072] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "fd55ddfc-f376-4e15-961e-6826f31a7890-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.304255] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.304431] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.306813] env[63355]: INFO nova.compute.manager [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Terminating instance [ 1037.309022] env[63355]: DEBUG nova.compute.manager [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1037.309334] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1037.310273] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1557f449-2860-4324-9dd4-2394a633837b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.323855] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.324959] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e4fb6d4-f139-423e-824a-53b1132f58b4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.334102] env[63355]: DEBUG oslo_vmware.api [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1037.334102] env[63355]: value = "task-1350066" [ 1037.334102] env[63355]: _type = "Task" [ 1037.334102] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.354671] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350063, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.355092] env[63355]: DEBUG oslo_vmware.api [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350066, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.410506] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c29210-c23e-4af6-8fea-fb190b3c560d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.422638] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1231829c-ab5c-4f29-a1c5-ec642efb5f0a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.469273] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f770d968-b547-4a9d-bcc2-170610dc6b6f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.480657] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186ae367-30c2-4c2c-a6d7-ee7b10ec78ed {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.498997] env[63355]: DEBUG nova.compute.provider_tree [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.672969] env[63355]: DEBUG nova.network.neutron [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Port 976ab0e0-5cea-4265-bc51-9b5cd29ed954 binding to destination host cpu-1 is already ACTIVE {{(pid=63355) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1037.738569] env[63355]: DEBUG nova.objects.instance [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'pci_requests' on Instance uuid da3cb83c-6368-49c0-9b11-0498221e3c0f {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.783938] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "50e7499c-a8c1-41a6-adac-36e6c3b92819" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.784256] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.852099] env[63355]: DEBUG oslo_vmware.api [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350066, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.855020] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350063, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.002809] env[63355]: DEBUG nova.scheduler.client.report [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.241718] env[63355]: DEBUG nova.objects.base [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1038.241962] env[63355]: DEBUG nova.network.neutron [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1038.287621] env[63355]: DEBUG nova.compute.manager [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1038.348654] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350063, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.015718} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.351648] env[63355]: INFO nova.virt.vmwareapi.ds_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac/OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac.vmdk to [datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383/c58460ed-ef70-483a-87f9-6152e2a0d383.vmdk. [ 1038.351878] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Cleaning up location [datastore2] OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1038.352078] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_2ad1d18d-e217-4bff-ae90-f4e63faee3ac {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.352366] env[63355]: DEBUG oslo_vmware.api [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350066, 'name': PowerOffVM_Task, 'duration_secs': 0.98795} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.353762] env[63355]: DEBUG nova.policy [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77243643ea724b72858a8682a2a054a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d47b671ea9c429391cbdae7e24adadf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1038.355674] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a22acad-4344-4164-a7c8-98ff16530a63 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.357403] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1038.358070] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1038.358070] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-34e33b7b-7680-4d68-88b5-5fef1a39c2ec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.365193] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1038.365193] env[63355]: value = "task-1350067" [ 1038.365193] env[63355]: _type = "Task" [ 1038.365193] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.374672] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350067, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.440801] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1038.441082] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1038.441307] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleting the datastore file [datastore2] fd55ddfc-f376-4e15-961e-6826f31a7890 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.441622] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2cbdba9b-d371-472d-aacd-799173546e81 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.449557] env[63355]: DEBUG oslo_vmware.api [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1038.449557] env[63355]: value = "task-1350069" [ 1038.449557] env[63355]: _type = "Task" [ 1038.449557] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.461107] env[63355]: DEBUG oslo_vmware.api [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350069, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.511332] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.285s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.511882] env[63355]: DEBUG nova.compute.manager [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1038.514791] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.039s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.516886] env[63355]: INFO nova.compute.claims [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1038.691259] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.691513] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.691694] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.812627] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.875773] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350067, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.083939} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.876063] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.876224] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383/c58460ed-ef70-483a-87f9-6152e2a0d383.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.876470] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383/c58460ed-ef70-483a-87f9-6152e2a0d383.vmdk to [datastore2] 7c153109-b814-4e11-b4f9-7b8cebb853d4/7c153109-b814-4e11-b4f9-7b8cebb853d4.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1038.876816] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb904609-e7ea-4ac1-b860-04c9d9f93cc8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.884876] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1038.884876] env[63355]: value = "task-1350070" [ 1038.884876] env[63355]: _type = "Task" [ 1038.884876] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.893619] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350070, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.959699] env[63355]: DEBUG oslo_vmware.api [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350069, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.021618] env[63355]: DEBUG nova.compute.utils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1039.024910] env[63355]: DEBUG nova.compute.manager [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1039.025102] env[63355]: DEBUG nova.network.neutron [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1039.071072] env[63355]: DEBUG nova.policy [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3721c4a73df54ae9b44110cb9e8590a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc547065748241e8ac7b6c499ddaea66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1039.354029] env[63355]: DEBUG nova.network.neutron [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Successfully created port: 93a72bba-8664-4305-a04f-e5e0449a27ff {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.397658] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350070, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.463095] env[63355]: DEBUG oslo_vmware.api [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350069, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.66172} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.463409] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.463605] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1039.463784] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1039.463966] env[63355]: INFO nova.compute.manager [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Took 2.15 seconds to destroy the instance on the hypervisor. [ 1039.464330] env[63355]: DEBUG oslo.service.loopingcall [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.464550] env[63355]: DEBUG nova.compute.manager [-] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.464709] env[63355]: DEBUG nova.network.neutron [-] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1039.526151] env[63355]: DEBUG nova.compute.manager [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1039.728671] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4dcc8ea-0ced-485b-b221-ce2908d6c38f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.744626] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf5f221-c17c-464f-bf84-08759a932348 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.783504] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c88640-2e18-40e4-9158-bcd5e0ab7692 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.795628] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1996b1-b6e1-4d8b-b16f-15ddd84825d9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.815062] env[63355]: DEBUG nova.compute.provider_tree [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.817425] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.817616] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.817794] env[63355]: DEBUG nova.network.neutron [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1039.906077] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350070, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.982998] env[63355]: DEBUG nova.compute.manager [req-464171b6-e2be-4f41-812e-8cf1fd873f3d req-4c8856b0-f53e-4342-a13b-74a51f01bbdc service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Received event network-vif-deleted-52d11eb4-329c-4989-9498-377fd46350cb {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.983225] env[63355]: INFO nova.compute.manager [req-464171b6-e2be-4f41-812e-8cf1fd873f3d req-4c8856b0-f53e-4342-a13b-74a51f01bbdc service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Neutron deleted interface 52d11eb4-329c-4989-9498-377fd46350cb; detaching it from the instance and deleting it from the info cache [ 1039.983416] env[63355]: DEBUG nova.network.neutron [req-464171b6-e2be-4f41-812e-8cf1fd873f3d req-4c8856b0-f53e-4342-a13b-74a51f01bbdc service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.174771] env[63355]: DEBUG nova.network.neutron [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Successfully updated port: 563020c5-4b41-4820-9807-82cf864fcade {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1040.320398] env[63355]: DEBUG nova.scheduler.client.report [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.327550] env[63355]: DEBUG nova.network.neutron [-] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.354798] env[63355]: DEBUG nova.compute.manager [req-979f0f22-5a86-4d21-87b3-62418d186c92 req-744e0652-71e2-49de-a147-42627b1abf78 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received event network-vif-plugged-563020c5-4b41-4820-9807-82cf864fcade {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.354798] env[63355]: DEBUG oslo_concurrency.lockutils [req-979f0f22-5a86-4d21-87b3-62418d186c92 req-744e0652-71e2-49de-a147-42627b1abf78 service nova] Acquiring lock "da3cb83c-6368-49c0-9b11-0498221e3c0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.354798] env[63355]: DEBUG oslo_concurrency.lockutils [req-979f0f22-5a86-4d21-87b3-62418d186c92 req-744e0652-71e2-49de-a147-42627b1abf78 service nova] Lock "da3cb83c-6368-49c0-9b11-0498221e3c0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.354798] env[63355]: DEBUG oslo_concurrency.lockutils [req-979f0f22-5a86-4d21-87b3-62418d186c92 req-744e0652-71e2-49de-a147-42627b1abf78 service nova] Lock "da3cb83c-6368-49c0-9b11-0498221e3c0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.354798] env[63355]: DEBUG nova.compute.manager [req-979f0f22-5a86-4d21-87b3-62418d186c92 req-744e0652-71e2-49de-a147-42627b1abf78 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] No waiting events found dispatching network-vif-plugged-563020c5-4b41-4820-9807-82cf864fcade {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1040.354798] env[63355]: WARNING nova.compute.manager [req-979f0f22-5a86-4d21-87b3-62418d186c92 req-744e0652-71e2-49de-a147-42627b1abf78 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received unexpected event network-vif-plugged-563020c5-4b41-4820-9807-82cf864fcade for instance with vm_state active and task_state None. [ 1040.401691] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350070, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.486649] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a376cc75-11dc-46b0-963c-107ffc6366bb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.501127] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db60b19-a321-44e9-8116-51a49e87d6cb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.533910] env[63355]: DEBUG nova.compute.manager [req-464171b6-e2be-4f41-812e-8cf1fd873f3d req-4c8856b0-f53e-4342-a13b-74a51f01bbdc service nova] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Detach interface failed, port_id=52d11eb4-329c-4989-9498-377fd46350cb, reason: Instance fd55ddfc-f376-4e15-961e-6826f31a7890 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1040.538506] env[63355]: DEBUG nova.compute.manager [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1040.567344] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.567581] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.567750] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.567949] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.568122] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.568277] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.568497] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.568662] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.568835] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.569035] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.569224] env[63355]: DEBUG nova.virt.hardware [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.570156] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7770e6-2795-48de-a0e3-70f81ba6b988 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.580777] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddae264f-bb7b-4081-8175-3052884f8ea2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.589444] env[63355]: DEBUG nova.network.neutron [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance_info_cache with network_info: [{"id": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "address": "fa:16:3e:b3:26:57", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap976ab0e0-5c", "ovs_interfaceid": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.678104] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.678360] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.678487] env[63355]: DEBUG nova.network.neutron [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1040.828725] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.829384] env[63355]: DEBUG nova.compute.manager [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1040.832096] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.020s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.833580] env[63355]: INFO nova.compute.claims [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1040.836050] env[63355]: INFO nova.compute.manager [-] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Took 1.37 seconds to deallocate network for instance. [ 1040.901149] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350070, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.007268] env[63355]: DEBUG nova.network.neutron [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Successfully updated port: 93a72bba-8664-4305-a04f-e5e0449a27ff {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1041.100700] env[63355]: DEBUG oslo_concurrency.lockutils [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.232676] env[63355]: WARNING nova.network.neutron [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] 943d3ecf-0c77-4c51-a997-b17d86259c34 already exists in list: networks containing: ['943d3ecf-0c77-4c51-a997-b17d86259c34']. ignoring it [ 1041.339090] env[63355]: DEBUG nova.compute.utils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.342271] env[63355]: DEBUG nova.compute.manager [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1041.342444] env[63355]: DEBUG nova.network.neutron [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1041.347460] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.400592] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350070, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.407103] env[63355]: DEBUG nova.policy [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'beb41983e69a44b2a108f552da5ef640', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfff1b9903264e5586119ebd3a3602de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1041.510034] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "refresh_cache-e271159e-14ed-4a88-a7b8-23d404d77985" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.510154] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "refresh_cache-e271159e-14ed-4a88-a7b8-23d404d77985" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.510298] env[63355]: DEBUG nova.network.neutron [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1041.551595] env[63355]: DEBUG nova.network.neutron [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [{"id": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "address": "fa:16:3e:08:9e:2a", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5a40b6-5d", "ovs_interfaceid": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "563020c5-4b41-4820-9807-82cf864fcade", "address": "fa:16:3e:de:c3:bd", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap563020c5-4b", "ovs_interfaceid": "563020c5-4b41-4820-9807-82cf864fcade", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.626113] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa672f30-f87d-4825-9d8a-f97ed819d6bf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.645713] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de8877d-35fc-464b-b4a0-5748bed5b737 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.654171] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance '537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5' progress to 83 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1041.670950] env[63355]: DEBUG nova.network.neutron [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Successfully created port: 6a28601f-1caa-4d3b-97c2-d913341d3bb9 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1041.850022] env[63355]: DEBUG nova.compute.manager [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1041.898979] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350070, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.530837} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.901243] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c58460ed-ef70-483a-87f9-6152e2a0d383/c58460ed-ef70-483a-87f9-6152e2a0d383.vmdk to [datastore2] 7c153109-b814-4e11-b4f9-7b8cebb853d4/7c153109-b814-4e11-b4f9-7b8cebb853d4.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1041.902437] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e63c6e-8c0f-4cc3-92ef-30044dfa0869 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.925282] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 7c153109-b814-4e11-b4f9-7b8cebb853d4/7c153109-b814-4e11-b4f9-7b8cebb853d4.vmdk or device None with type streamOptimized {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.928127] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d563ded8-7cf7-4a1c-9ac0-4b19603ca05e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.948601] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1041.948601] env[63355]: value = "task-1350071" [ 1041.948601] env[63355]: _type = "Task" [ 1041.948601] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.958930] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350071, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.005876] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7382be-fcb6-4aaf-a240-c203785b0040 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.010667] env[63355]: DEBUG nova.compute.manager [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Received event network-vif-plugged-93a72bba-8664-4305-a04f-e5e0449a27ff {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.010926] env[63355]: DEBUG oslo_concurrency.lockutils [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] Acquiring lock "e271159e-14ed-4a88-a7b8-23d404d77985-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.011101] env[63355]: DEBUG oslo_concurrency.lockutils [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] Lock "e271159e-14ed-4a88-a7b8-23d404d77985-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.011274] env[63355]: DEBUG oslo_concurrency.lockutils [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] Lock "e271159e-14ed-4a88-a7b8-23d404d77985-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.011443] env[63355]: DEBUG nova.compute.manager [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] No waiting events found dispatching network-vif-plugged-93a72bba-8664-4305-a04f-e5e0449a27ff {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1042.011605] env[63355]: WARNING nova.compute.manager [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Received unexpected event network-vif-plugged-93a72bba-8664-4305-a04f-e5e0449a27ff for instance with vm_state building and task_state spawning. [ 1042.011767] env[63355]: DEBUG nova.compute.manager [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Received event network-changed-93a72bba-8664-4305-a04f-e5e0449a27ff {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.011922] env[63355]: DEBUG nova.compute.manager [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Refreshing instance network info cache due to event network-changed-93a72bba-8664-4305-a04f-e5e0449a27ff. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1042.012106] env[63355]: DEBUG oslo_concurrency.lockutils [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] Acquiring lock "refresh_cache-e271159e-14ed-4a88-a7b8-23d404d77985" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.017668] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c381352e-29fc-4532-9644-87d9b7863c99 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.049034] env[63355]: DEBUG nova.network.neutron [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1042.051277] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0371004e-8961-4db4-955f-19f54588ef70 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.053900] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.054458] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.054620] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.055625] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb510546-3c52-48f8-bcf0-135870417667 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.074698] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b9fc70-bf16-4935-8b2f-0d72184b3323 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.079383] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.079635] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.079796] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.079980] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.080148] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.080297] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.080503] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.080664] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.080833] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.081007] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.081194] env[63355]: DEBUG nova.virt.hardware [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.087523] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Reconfiguring VM to attach interface {{(pid=63355) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1042.089817] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fcccfc5-2082-4d5b-bbea-bd85679e0d5a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.112628] env[63355]: DEBUG nova.compute.provider_tree [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.114863] env[63355]: DEBUG oslo_vmware.api [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1042.114863] env[63355]: value = "task-1350072" [ 1042.114863] env[63355]: _type = "Task" [ 1042.114863] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.123674] env[63355]: DEBUG oslo_vmware.api [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350072, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.170147] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4ba8-4214-4ae6-ba8b-f6013c9ff4bc tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance '537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5' progress to 100 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1042.402755] env[63355]: DEBUG nova.network.neutron [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Updating instance_info_cache with network_info: [{"id": "93a72bba-8664-4305-a04f-e5e0449a27ff", "address": "fa:16:3e:f8:27:2c", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93a72bba-86", "ovs_interfaceid": "93a72bba-8664-4305-a04f-e5e0449a27ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.436269] env[63355]: DEBUG nova.compute.manager [req-800d0270-8d43-44a0-906f-0ddf8cae30e0 req-4eb2d7bc-7249-4c80-bf7c-f141da279233 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received event network-changed-563020c5-4b41-4820-9807-82cf864fcade {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.436517] env[63355]: DEBUG nova.compute.manager [req-800d0270-8d43-44a0-906f-0ddf8cae30e0 req-4eb2d7bc-7249-4c80-bf7c-f141da279233 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing instance network info cache due to event network-changed-563020c5-4b41-4820-9807-82cf864fcade. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1042.436800] env[63355]: DEBUG oslo_concurrency.lockutils [req-800d0270-8d43-44a0-906f-0ddf8cae30e0 req-4eb2d7bc-7249-4c80-bf7c-f141da279233 service nova] Acquiring lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.437040] env[63355]: DEBUG oslo_concurrency.lockutils [req-800d0270-8d43-44a0-906f-0ddf8cae30e0 req-4eb2d7bc-7249-4c80-bf7c-f141da279233 service nova] Acquired lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.437185] env[63355]: DEBUG nova.network.neutron [req-800d0270-8d43-44a0-906f-0ddf8cae30e0 req-4eb2d7bc-7249-4c80-bf7c-f141da279233 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing network info cache for port 563020c5-4b41-4820-9807-82cf864fcade {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1042.459067] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350071, 'name': ReconfigVM_Task, 'duration_secs': 0.332618} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.459067] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 7c153109-b814-4e11-b4f9-7b8cebb853d4/7c153109-b814-4e11-b4f9-7b8cebb853d4.vmdk or device None with type streamOptimized {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.459747] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1915bdac-a514-4cf6-a2a2-910d9e0c4ff1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.465729] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1042.465729] env[63355]: value = "task-1350073" [ 1042.465729] env[63355]: _type = "Task" [ 1042.465729] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.476464] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350073, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.617395] env[63355]: DEBUG nova.scheduler.client.report [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.638933] env[63355]: DEBUG oslo_vmware.api [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350072, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.860686] env[63355]: DEBUG nova.compute.manager [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1042.886231] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.886610] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.886874] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.887163] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.887385] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.887602] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.887903] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.888148] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.888414] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.888622] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.888804] env[63355]: DEBUG nova.virt.hardware [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.889682] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759218d7-4d16-4cfd-8e31-97e0233a2c03 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.898416] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04af64b-185c-47b5-91d4-774da028ad6b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.911990] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "refresh_cache-e271159e-14ed-4a88-a7b8-23d404d77985" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.912307] env[63355]: DEBUG nova.compute.manager [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Instance network_info: |[{"id": "93a72bba-8664-4305-a04f-e5e0449a27ff", "address": "fa:16:3e:f8:27:2c", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93a72bba-86", "ovs_interfaceid": "93a72bba-8664-4305-a04f-e5e0449a27ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1042.912793] env[63355]: DEBUG oslo_concurrency.lockutils [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] Acquired lock "refresh_cache-e271159e-14ed-4a88-a7b8-23d404d77985" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.912980] env[63355]: DEBUG nova.network.neutron [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Refreshing network info cache for port 93a72bba-8664-4305-a04f-e5e0449a27ff {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1042.914190] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:27:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '93a72bba-8664-4305-a04f-e5e0449a27ff', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.921693] env[63355]: DEBUG oslo.service.loopingcall [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.924620] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1042.925114] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1100cb28-a0e8-4e39-8f5c-7192add6da91 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.948629] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.948629] env[63355]: value = "task-1350074" [ 1042.948629] env[63355]: _type = "Task" [ 1042.948629] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.958514] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350074, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.975183] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350073, 'name': Rename_Task, 'duration_secs': 0.156032} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.975409] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1042.975665] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e14f6cf-9da7-4149-a9d6-fa660bac1d81 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.982053] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1042.982053] env[63355]: value = "task-1350075" [ 1042.982053] env[63355]: _type = "Task" [ 1042.982053] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.990711] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350075, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.132982] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.301s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.133517] env[63355]: DEBUG nova.compute.manager [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1043.135986] env[63355]: DEBUG oslo_vmware.api [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350072, 'name': ReconfigVM_Task, 'duration_secs': 0.721742} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.136293] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.789s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.136504] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.138716] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.138909] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Reconfigured VM to attach interface {{(pid=63355) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1043.156224] env[63355]: DEBUG nova.network.neutron [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Successfully updated port: 6a28601f-1caa-4d3b-97c2-d913341d3bb9 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1043.165813] env[63355]: INFO nova.scheduler.client.report [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted allocations for instance fd55ddfc-f376-4e15-961e-6826f31a7890 [ 1043.304453] env[63355]: DEBUG nova.network.neutron [req-800d0270-8d43-44a0-906f-0ddf8cae30e0 req-4eb2d7bc-7249-4c80-bf7c-f141da279233 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updated VIF entry in instance network info cache for port 563020c5-4b41-4820-9807-82cf864fcade. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1043.304987] env[63355]: DEBUG nova.network.neutron [req-800d0270-8d43-44a0-906f-0ddf8cae30e0 req-4eb2d7bc-7249-4c80-bf7c-f141da279233 service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [{"id": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "address": "fa:16:3e:08:9e:2a", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5a40b6-5d", "ovs_interfaceid": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "563020c5-4b41-4820-9807-82cf864fcade", "address": "fa:16:3e:de:c3:bd", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap563020c5-4b", "ovs_interfaceid": "563020c5-4b41-4820-9807-82cf864fcade", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.307745] env[63355]: DEBUG nova.network.neutron [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Updated VIF entry in instance network info cache for port 93a72bba-8664-4305-a04f-e5e0449a27ff. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1043.308097] env[63355]: DEBUG nova.network.neutron [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Updating instance_info_cache with network_info: [{"id": "93a72bba-8664-4305-a04f-e5e0449a27ff", "address": "fa:16:3e:f8:27:2c", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93a72bba-86", "ovs_interfaceid": "93a72bba-8664-4305-a04f-e5e0449a27ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.460146] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350074, 'name': CreateVM_Task, 'duration_secs': 0.399635} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.461051] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1043.461051] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.461241] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.461577] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.461893] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b556926-f642-4756-b795-5feb4dad58a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.467388] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1043.467388] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5278a3a6-562c-05e3-baa0-9cf6d9e6cacf" [ 1043.467388] env[63355]: _type = "Task" [ 1043.467388] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.476567] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5278a3a6-562c-05e3-baa0-9cf6d9e6cacf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.492646] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350075, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.643910] env[63355]: DEBUG nova.compute.utils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.645958] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97efa47f-1244-486c-a29d-f2e18878f0a9 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-da3cb83c-6368-49c0-9b11-0498221e3c0f-563020c5-4b41-4820-9807-82cf864fcade" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.570s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.646883] env[63355]: DEBUG nova.compute.manager [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1043.647060] env[63355]: DEBUG nova.network.neutron [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1043.658440] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-e398ee67-eb02-4256-9120-06111f110692" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.658595] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-e398ee67-eb02-4256-9120-06111f110692" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.658745] env[63355]: DEBUG nova.network.neutron [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1043.675907] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b87a3af8-a6c3-4a62-ba2a-5fb62a9e0519 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "fd55ddfc-f376-4e15-961e-6826f31a7890" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.372s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.694090] env[63355]: DEBUG nova.policy [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bf70ebeeb3a4327ada51d18f5651476', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce0682a99ac94aeea463c961b84e6b58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1043.807462] env[63355]: DEBUG oslo_concurrency.lockutils [req-800d0270-8d43-44a0-906f-0ddf8cae30e0 req-4eb2d7bc-7249-4c80-bf7c-f141da279233 service nova] Releasing lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.810234] env[63355]: DEBUG oslo_concurrency.lockutils [req-462cc260-b2ff-4a66-bdf7-5a2e9354c994 req-8d5689f7-a75c-4b2b-9d48-8e2cae1be36c service nova] Releasing lock "refresh_cache-e271159e-14ed-4a88-a7b8-23d404d77985" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.936238] env[63355]: DEBUG nova.network.neutron [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Successfully created port: ba0aaec1-3b3e-4594-bf65-aa18843e439d {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1043.978149] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5278a3a6-562c-05e3-baa0-9cf6d9e6cacf, 'name': SearchDatastore_Task, 'duration_secs': 0.012228} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.978454] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.978687] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.978930] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.979109] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.979296] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.979558] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d39af58-bfcf-4036-aa07-7e53194a126a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.989415] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.989603] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1043.993164] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02570256-34ef-48c7-90bf-893f4d75282b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.995331] env[63355]: DEBUG oslo_vmware.api [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350075, 'name': PowerOnVM_Task, 'duration_secs': 0.521974} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.995588] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1044.000486] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1044.000486] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52cea7b8-cc43-b3f2-1d04-258ed61358fd" [ 1044.000486] env[63355]: _type = "Task" [ 1044.000486] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.008223] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52cea7b8-cc43-b3f2-1d04-258ed61358fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.080285] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.080463] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.080855] env[63355]: DEBUG nova.compute.manager [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Going to confirm migration 2 {{(pid=63355) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1044.103428] env[63355]: DEBUG nova.compute.manager [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1044.104336] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8791759-9461-4728-8abc-0e0dca938efe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.151853] env[63355]: DEBUG nova.compute.manager [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1044.195168] env[63355]: DEBUG nova.network.neutron [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1044.322019] env[63355]: DEBUG nova.network.neutron [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Updating instance_info_cache with network_info: [{"id": "6a28601f-1caa-4d3b-97c2-d913341d3bb9", "address": "fa:16:3e:89:3e:16", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a28601f-1c", "ovs_interfaceid": "6a28601f-1caa-4d3b-97c2-d913341d3bb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.511213] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52cea7b8-cc43-b3f2-1d04-258ed61358fd, 'name': SearchDatastore_Task, 'duration_secs': 0.015317} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.512872] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca705bd0-6222-496c-aee9-053aaf253b63 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.517306] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1044.517306] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e7e978-8606-c067-6fbd-0244a1c4c0d2" [ 1044.517306] env[63355]: _type = "Task" [ 1044.517306] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.525341] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e7e978-8606-c067-6fbd-0244a1c4c0d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.528023] env[63355]: DEBUG nova.compute.manager [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Received event network-vif-plugged-6a28601f-1caa-4d3b-97c2-d913341d3bb9 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.528327] env[63355]: DEBUG oslo_concurrency.lockutils [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] Acquiring lock "e398ee67-eb02-4256-9120-06111f110692-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.528630] env[63355]: DEBUG oslo_concurrency.lockutils [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] Lock "e398ee67-eb02-4256-9120-06111f110692-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.528862] env[63355]: DEBUG oslo_concurrency.lockutils [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] Lock "e398ee67-eb02-4256-9120-06111f110692-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.529093] env[63355]: DEBUG nova.compute.manager [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] No waiting events found dispatching network-vif-plugged-6a28601f-1caa-4d3b-97c2-d913341d3bb9 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1044.529336] env[63355]: WARNING nova.compute.manager [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Received unexpected event network-vif-plugged-6a28601f-1caa-4d3b-97c2-d913341d3bb9 for instance with vm_state building and task_state spawning. [ 1044.529516] env[63355]: DEBUG nova.compute.manager [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Received event network-changed-6a28601f-1caa-4d3b-97c2-d913341d3bb9 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.529677] env[63355]: DEBUG nova.compute.manager [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Refreshing instance network info cache due to event network-changed-6a28601f-1caa-4d3b-97c2-d913341d3bb9. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1044.529854] env[63355]: DEBUG oslo_concurrency.lockutils [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] Acquiring lock "refresh_cache-e398ee67-eb02-4256-9120-06111f110692" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.623027] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ecfd91ed-21fa-49e2-a45b-200e34fd6598 tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 25.234s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.663516] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.663709] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.663890] env[63355]: DEBUG nova.network.neutron [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1044.664095] env[63355]: DEBUG nova.objects.instance [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'info_cache' on Instance uuid 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.824585] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-e398ee67-eb02-4256-9120-06111f110692" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.825242] env[63355]: DEBUG nova.compute.manager [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Instance network_info: |[{"id": "6a28601f-1caa-4d3b-97c2-d913341d3bb9", "address": "fa:16:3e:89:3e:16", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a28601f-1c", "ovs_interfaceid": "6a28601f-1caa-4d3b-97c2-d913341d3bb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1044.825394] env[63355]: DEBUG oslo_concurrency.lockutils [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] Acquired lock "refresh_cache-e398ee67-eb02-4256-9120-06111f110692" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.825484] env[63355]: DEBUG nova.network.neutron [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Refreshing network info cache for port 6a28601f-1caa-4d3b-97c2-d913341d3bb9 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1044.826802] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:3e:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '52358fcc-0d9f-45dd-8c75-db533fd992c3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6a28601f-1caa-4d3b-97c2-d913341d3bb9', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.835636] env[63355]: DEBUG oslo.service.loopingcall [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.838695] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e398ee67-eb02-4256-9120-06111f110692] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1044.839227] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a8bcc24d-ead2-4b13-b94f-0824abd16ab5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.860552] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.860552] env[63355]: value = "task-1350076" [ 1044.860552] env[63355]: _type = "Task" [ 1044.860552] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.869358] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350076, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.027964] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e7e978-8606-c067-6fbd-0244a1c4c0d2, 'name': SearchDatastore_Task, 'duration_secs': 0.011029} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.028253] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.028543] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] e271159e-14ed-4a88-a7b8-23d404d77985/e271159e-14ed-4a88-a7b8-23d404d77985.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1045.028788] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65d43c20-d9e7-4f7e-9808-74fa282ec1ab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.035098] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1045.035098] env[63355]: value = "task-1350077" [ 1045.035098] env[63355]: _type = "Task" [ 1045.035098] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.042928] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350077, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.047387] env[63355]: DEBUG nova.network.neutron [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Updated VIF entry in instance network info cache for port 6a28601f-1caa-4d3b-97c2-d913341d3bb9. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1045.047730] env[63355]: DEBUG nova.network.neutron [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Updating instance_info_cache with network_info: [{"id": "6a28601f-1caa-4d3b-97c2-d913341d3bb9", "address": "fa:16:3e:89:3e:16", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a28601f-1c", "ovs_interfaceid": "6a28601f-1caa-4d3b-97c2-d913341d3bb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.161109] env[63355]: DEBUG nova.compute.manager [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1045.187485] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1045.187855] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1045.188089] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.188339] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1045.188521] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.188704] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1045.188960] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1045.189169] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1045.189381] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1045.189572] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1045.189784] env[63355]: DEBUG nova.virt.hardware [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.191050] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac992c9-66d7-4b58-8038-d7816dfad66e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.199976] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5412a34f-2b5d-44f8-a5aa-e607e7e04ef6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.246355] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "interface-da3cb83c-6368-49c0-9b11-0498221e3c0f-563020c5-4b41-4820-9807-82cf864fcade" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.246644] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-da3cb83c-6368-49c0-9b11-0498221e3c0f-563020c5-4b41-4820-9807-82cf864fcade" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.373183] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350076, 'name': CreateVM_Task, 'duration_secs': 0.338568} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.374664] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e398ee67-eb02-4256-9120-06111f110692] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1045.375835] env[63355]: DEBUG nova.compute.manager [req-570e26ca-2189-4d23-b699-d527a3c6fa8f req-d06f842d-6eab-4567-b34a-afbf0519cb27 service nova] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Received event network-vif-plugged-ba0aaec1-3b3e-4594-bf65-aa18843e439d {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.376140] env[63355]: DEBUG oslo_concurrency.lockutils [req-570e26ca-2189-4d23-b699-d527a3c6fa8f req-d06f842d-6eab-4567-b34a-afbf0519cb27 service nova] Acquiring lock "50e7499c-a8c1-41a6-adac-36e6c3b92819-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.376271] env[63355]: DEBUG oslo_concurrency.lockutils [req-570e26ca-2189-4d23-b699-d527a3c6fa8f req-d06f842d-6eab-4567-b34a-afbf0519cb27 service nova] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.376445] env[63355]: DEBUG oslo_concurrency.lockutils [req-570e26ca-2189-4d23-b699-d527a3c6fa8f req-d06f842d-6eab-4567-b34a-afbf0519cb27 service nova] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.376615] env[63355]: DEBUG nova.compute.manager [req-570e26ca-2189-4d23-b699-d527a3c6fa8f req-d06f842d-6eab-4567-b34a-afbf0519cb27 service nova] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] No waiting events found dispatching network-vif-plugged-ba0aaec1-3b3e-4594-bf65-aa18843e439d {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1045.376871] env[63355]: WARNING nova.compute.manager [req-570e26ca-2189-4d23-b699-d527a3c6fa8f req-d06f842d-6eab-4567-b34a-afbf0519cb27 service nova] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Received unexpected event network-vif-plugged-ba0aaec1-3b3e-4594-bf65-aa18843e439d for instance with vm_state building and task_state spawning. [ 1045.377596] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.377796] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.378239] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1045.378846] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-686507b6-59ec-4257-8566-d59a90f6a1cb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.387071] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1045.387071] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521d4b7c-cf40-2fa3-92f6-258a832dd3e7" [ 1045.387071] env[63355]: _type = "Task" [ 1045.387071] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.397343] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521d4b7c-cf40-2fa3-92f6-258a832dd3e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.481137] env[63355]: DEBUG nova.network.neutron [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Successfully updated port: ba0aaec1-3b3e-4594-bf65-aa18843e439d {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1045.545019] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350077, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476053} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.545310] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] e271159e-14ed-4a88-a7b8-23d404d77985/e271159e-14ed-4a88-a7b8-23d404d77985.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1045.545533] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.545793] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6161416-239f-4848-9931-a1c5fe4efe8d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.550496] env[63355]: DEBUG oslo_concurrency.lockutils [req-128764d7-d812-4615-86b0-2df52a253d00 req-b858f98a-1b4a-475a-a2e0-9e8837852673 service nova] Releasing lock "refresh_cache-e398ee67-eb02-4256-9120-06111f110692" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.552727] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1045.552727] env[63355]: value = "task-1350078" [ 1045.552727] env[63355]: _type = "Task" [ 1045.552727] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.561973] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350078, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.750319] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.750555] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.751804] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802f3eb7-c2d1-4f86-a408-bb6439c0f751 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.771287] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762e671e-ff67-4391-9a51-e35e1c0e1b98 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.797373] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Reconfiguring VM to detach interface {{(pid=63355) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1045.799824] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-362995e9-52e5-47d7-b4da-8949e487c5b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.818860] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1045.818860] env[63355]: value = "task-1350079" [ 1045.818860] env[63355]: _type = "Task" [ 1045.818860] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.828772] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.899531] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]521d4b7c-cf40-2fa3-92f6-258a832dd3e7, 'name': SearchDatastore_Task, 'duration_secs': 0.039199} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.899838] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.900084] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1045.900320] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.900469] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.900645] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1045.900913] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5242c0d-556d-4d54-9f29-9ef7bcc1c048 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.908552] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1045.909526] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1045.909526] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78b5cb2e-0b88-40fa-9263-053a19c48b16 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.914250] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1045.914250] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b1b538-0d10-ea8c-66a0-05e212c94348" [ 1045.914250] env[63355]: _type = "Task" [ 1045.914250] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.921466] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b1b538-0d10-ea8c-66a0-05e212c94348, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.924981] env[63355]: DEBUG nova.network.neutron [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance_info_cache with network_info: [{"id": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "address": "fa:16:3e:b3:26:57", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap976ab0e0-5c", "ovs_interfaceid": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.983437] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.983564] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.983712] env[63355]: DEBUG nova.network.neutron [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.063057] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350078, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065508} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.063333] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1046.064137] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ec801c-b435-475f-9ac4-4a19d9395bdd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.086116] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] e271159e-14ed-4a88-a7b8-23d404d77985/e271159e-14ed-4a88-a7b8-23d404d77985.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.086436] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01fff39a-7755-437e-b68e-df53a09041a0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.106046] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1046.106046] env[63355]: value = "task-1350080" [ 1046.106046] env[63355]: _type = "Task" [ 1046.106046] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.113918] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350080, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.329127] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.424019] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b1b538-0d10-ea8c-66a0-05e212c94348, 'name': SearchDatastore_Task, 'duration_secs': 0.008523} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.424857] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04bc708d-721d-4d9d-acf6-6c2c9d6805f2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.427249] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.427538] env[63355]: DEBUG nova.objects.instance [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'migration_context' on Instance uuid 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.431873] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1046.431873] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528d77d2-65b8-5d43-1c71-22619a9cb046" [ 1046.431873] env[63355]: _type = "Task" [ 1046.431873] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.440687] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528d77d2-65b8-5d43-1c71-22619a9cb046, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.516384] env[63355]: DEBUG nova.network.neutron [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1046.618810] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350080, 'name': ReconfigVM_Task, 'duration_secs': 0.270384} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.619118] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Reconfigured VM instance instance-0000005e to attach disk [datastore2] e271159e-14ed-4a88-a7b8-23d404d77985/e271159e-14ed-4a88-a7b8-23d404d77985.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.619778] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a11c863f-e061-465f-8ff4-7d399e4a5eb8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.625850] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1046.625850] env[63355]: value = "task-1350081" [ 1046.625850] env[63355]: _type = "Task" [ 1046.625850] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.633687] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350081, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.647292] env[63355]: DEBUG nova.network.neutron [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance_info_cache with network_info: [{"id": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "address": "fa:16:3e:8a:f8:58", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba0aaec1-3b", "ovs_interfaceid": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.831140] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.930648] env[63355]: DEBUG nova.objects.base [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Object Instance<537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5> lazy-loaded attributes: info_cache,migration_context {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1046.931608] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4716c071-dded-4103-bfbb-e6a5b669a4f6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.943361] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528d77d2-65b8-5d43-1c71-22619a9cb046, 'name': SearchDatastore_Task, 'duration_secs': 0.00959} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.956224] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.956528] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] e398ee67-eb02-4256-9120-06111f110692/e398ee67-eb02-4256-9120-06111f110692.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1046.957411] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34b52eba-11ed-410e-a58f-7a22a4adac83 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.959308] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5e57897-7ce3-4a3d-adb3-a996407386f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.965549] env[63355]: DEBUG oslo_vmware.api [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1046.965549] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52cdb278-14ce-054e-eeef-233e78a23c5a" [ 1046.965549] env[63355]: _type = "Task" [ 1046.965549] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.966783] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1046.966783] env[63355]: value = "task-1350082" [ 1046.966783] env[63355]: _type = "Task" [ 1046.966783] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.976138] env[63355]: DEBUG oslo_vmware.api [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52cdb278-14ce-054e-eeef-233e78a23c5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.978885] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350082, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.135601] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350081, 'name': Rename_Task, 'duration_secs': 0.145141} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.135882] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1047.136171] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b9ee2ee-04b3-417f-af5c-efadf419350f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.143013] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1047.143013] env[63355]: value = "task-1350083" [ 1047.143013] env[63355]: _type = "Task" [ 1047.143013] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.149987] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.150303] env[63355]: DEBUG nova.compute.manager [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Instance network_info: |[{"id": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "address": "fa:16:3e:8a:f8:58", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba0aaec1-3b", "ovs_interfaceid": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1047.150565] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350083, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.150901] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:f8:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '130387c4-e4ec-4d95-8e9d-bb079baabad8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba0aaec1-3b3e-4594-bf65-aa18843e439d', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1047.158289] env[63355]: DEBUG oslo.service.loopingcall [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.158490] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1047.158696] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-040fc2ce-5f26-455b-af09-5f0280d2e12e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.178468] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1047.178468] env[63355]: value = "task-1350084" [ 1047.178468] env[63355]: _type = "Task" [ 1047.178468] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.186391] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350084, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.331697] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.405101] env[63355]: DEBUG nova.compute.manager [req-eaf3b58d-50f2-463e-a95d-8e3db32778f4 req-286ecb4a-9224-485c-bb7e-77ce27adf142 service nova] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Received event network-changed-ba0aaec1-3b3e-4594-bf65-aa18843e439d {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.405480] env[63355]: DEBUG nova.compute.manager [req-eaf3b58d-50f2-463e-a95d-8e3db32778f4 req-286ecb4a-9224-485c-bb7e-77ce27adf142 service nova] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Refreshing instance network info cache due to event network-changed-ba0aaec1-3b3e-4594-bf65-aa18843e439d. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1047.405777] env[63355]: DEBUG oslo_concurrency.lockutils [req-eaf3b58d-50f2-463e-a95d-8e3db32778f4 req-286ecb4a-9224-485c-bb7e-77ce27adf142 service nova] Acquiring lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.405981] env[63355]: DEBUG oslo_concurrency.lockutils [req-eaf3b58d-50f2-463e-a95d-8e3db32778f4 req-286ecb4a-9224-485c-bb7e-77ce27adf142 service nova] Acquired lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.406219] env[63355]: DEBUG nova.network.neutron [req-eaf3b58d-50f2-463e-a95d-8e3db32778f4 req-286ecb4a-9224-485c-bb7e-77ce27adf142 service nova] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Refreshing network info cache for port ba0aaec1-3b3e-4594-bf65-aa18843e439d {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1047.481147] env[63355]: DEBUG oslo_vmware.api [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52cdb278-14ce-054e-eeef-233e78a23c5a, 'name': SearchDatastore_Task, 'duration_secs': 0.010018} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.484520] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.484885] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.486609] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350082, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.653745] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350083, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.687656] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350084, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.831684] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.981310] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350082, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.536028} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.983284] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] e398ee67-eb02-4256-9120-06111f110692/e398ee67-eb02-4256-9120-06111f110692.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1047.983503] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1047.983759] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea2db082-5b8d-452b-a014-a0cdceacd7f4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.991970] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1047.991970] env[63355]: value = "task-1350085" [ 1047.991970] env[63355]: _type = "Task" [ 1047.991970] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.998757] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350085, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.105921] env[63355]: DEBUG nova.network.neutron [req-eaf3b58d-50f2-463e-a95d-8e3db32778f4 req-286ecb4a-9224-485c-bb7e-77ce27adf142 service nova] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updated VIF entry in instance network info cache for port ba0aaec1-3b3e-4594-bf65-aa18843e439d. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.106340] env[63355]: DEBUG nova.network.neutron [req-eaf3b58d-50f2-463e-a95d-8e3db32778f4 req-286ecb4a-9224-485c-bb7e-77ce27adf142 service nova] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance_info_cache with network_info: [{"id": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "address": "fa:16:3e:8a:f8:58", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba0aaec1-3b", "ovs_interfaceid": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.116536] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86890e5c-c2b7-4a1d-8bf7-9bd4c1d64282 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.124588] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19db33f6-0078-4082-b2a6-488ed9f22dc7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.156228] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9e9ef0-1123-45f2-9265-9d7a3a351fb6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.163817] env[63355]: DEBUG oslo_vmware.api [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350083, 'name': PowerOnVM_Task, 'duration_secs': 0.514071} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.165807] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.166065] env[63355]: INFO nova.compute.manager [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Took 7.63 seconds to spawn the instance on the hypervisor. [ 1048.166254] env[63355]: DEBUG nova.compute.manager [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.167037] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef05420-d854-4a19-8c24-7e09f3848e01 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.170175] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-058262f7-851d-400d-b415-dae731b2cfe6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.189291] env[63355]: DEBUG nova.compute.provider_tree [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.197696] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350084, 'name': CreateVM_Task, 'duration_secs': 0.563661} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.198413] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1048.199072] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.199244] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.199555] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1048.200029] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2135ebac-3b06-401f-9d2f-855ea3466a81 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.204230] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1048.204230] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52951c24-fb33-53e6-3490-5edb3ba87cc7" [ 1048.204230] env[63355]: _type = "Task" [ 1048.204230] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.211840] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52951c24-fb33-53e6-3490-5edb3ba87cc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.331924] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.500442] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350085, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065622} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.500663] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1048.501488] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c11a9f4-b280-4d25-afe8-bfd8feda56ec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.522938] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] e398ee67-eb02-4256-9120-06111f110692/e398ee67-eb02-4256-9120-06111f110692.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.523183] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3fc700d-a274-46bd-8961-dc01e9a94f27 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.542180] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1048.542180] env[63355]: value = "task-1350086" [ 1048.542180] env[63355]: _type = "Task" [ 1048.542180] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.549376] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350086, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.609414] env[63355]: DEBUG oslo_concurrency.lockutils [req-eaf3b58d-50f2-463e-a95d-8e3db32778f4 req-286ecb4a-9224-485c-bb7e-77ce27adf142 service nova] Releasing lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.696050] env[63355]: DEBUG nova.scheduler.client.report [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.700272] env[63355]: INFO nova.compute.manager [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Took 13.97 seconds to build instance. [ 1048.715324] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52951c24-fb33-53e6-3490-5edb3ba87cc7, 'name': SearchDatastore_Task, 'duration_secs': 0.009763} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.715607] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.715831] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1048.716082] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.716238] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.716421] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1048.716907] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ccdc177-221e-4e49-8221-5b154a293247 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.725092] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1048.725276] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1048.726057] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11d05cef-7b9c-418c-ac47-fb2da18a4b84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.731476] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1048.731476] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52876ec4-833c-26b1-7eb1-98056928acb6" [ 1048.731476] env[63355]: _type = "Task" [ 1048.731476] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.739467] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52876ec4-833c-26b1-7eb1-98056928acb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.833308] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.052744] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.203918] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c33870cc-1563-45a8-b88c-5f691912566f tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "e271159e-14ed-4a88-a7b8-23d404d77985" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.485s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.242102] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52876ec4-833c-26b1-7eb1-98056928acb6, 'name': SearchDatastore_Task, 'duration_secs': 0.010614} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.242918] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-477df365-d00c-48e6-897c-5ae932688102 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.248475] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1049.248475] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]520f77ff-dd0b-9726-3a17-4d32c8c7607c" [ 1049.248475] env[63355]: _type = "Task" [ 1049.248475] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.257806] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]520f77ff-dd0b-9726-3a17-4d32c8c7607c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.333079] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.551971] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350086, 'name': ReconfigVM_Task, 'duration_secs': 0.620426} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.552332] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Reconfigured VM instance instance-0000005f to attach disk [datastore1] e398ee67-eb02-4256-9120-06111f110692/e398ee67-eb02-4256-9120-06111f110692.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.552941] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2b21f519-0e82-4458-a67a-2bea9681ba4e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.558770] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1049.558770] env[63355]: value = "task-1350087" [ 1049.558770] env[63355]: _type = "Task" [ 1049.558770] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.565911] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350087, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.706663] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.222s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.707018] env[63355]: DEBUG nova.compute.manager [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=63355) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1049.758700] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]520f77ff-dd0b-9726-3a17-4d32c8c7607c, 'name': SearchDatastore_Task, 'duration_secs': 0.010616} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.759038] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.759349] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 50e7499c-a8c1-41a6-adac-36e6c3b92819/50e7499c-a8c1-41a6-adac-36e6c3b92819.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1049.759614] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb35d8ab-c861-4b06-b1de-2698718716db {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.766942] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1049.766942] env[63355]: value = "task-1350088" [ 1049.766942] env[63355]: _type = "Task" [ 1049.766942] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.774259] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350088, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.833976] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.992394] env[63355]: DEBUG oslo_concurrency.lockutils [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "e271159e-14ed-4a88-a7b8-23d404d77985" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.993098] env[63355]: DEBUG oslo_concurrency.lockutils [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "e271159e-14ed-4a88-a7b8-23d404d77985" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.993098] env[63355]: DEBUG oslo_concurrency.lockutils [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "e271159e-14ed-4a88-a7b8-23d404d77985-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.993098] env[63355]: DEBUG oslo_concurrency.lockutils [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "e271159e-14ed-4a88-a7b8-23d404d77985-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.993425] env[63355]: DEBUG oslo_concurrency.lockutils [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "e271159e-14ed-4a88-a7b8-23d404d77985-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.995722] env[63355]: INFO nova.compute.manager [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Terminating instance [ 1049.997966] env[63355]: DEBUG nova.compute.manager [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1049.998171] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1049.999011] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e22746-885b-4361-b062-1f0b22f8dda1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.007036] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1050.007317] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-115f32f6-5477-4803-a3cb-6c62db532dfc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.014404] env[63355]: DEBUG oslo_vmware.api [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1050.014404] env[63355]: value = "task-1350089" [ 1050.014404] env[63355]: _type = "Task" [ 1050.014404] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.022877] env[63355]: DEBUG oslo_vmware.api [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350089, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.068703] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350087, 'name': Rename_Task, 'duration_secs': 0.138729} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.069022] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.069353] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9de87b5a-a830-44a8-ba4c-67daf1ec1a5a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.076338] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1050.076338] env[63355]: value = "task-1350090" [ 1050.076338] env[63355]: _type = "Task" [ 1050.076338] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.085313] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350090, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.271626] env[63355]: INFO nova.scheduler.client.report [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleted allocation for migration 0d5cde65-ecd9-4f5b-b4bf-dfd872c38a3f [ 1050.279236] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350088, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.335922] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.525101] env[63355]: DEBUG oslo_vmware.api [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350089, 'name': PowerOffVM_Task, 'duration_secs': 0.201556} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.525419] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1050.525585] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1050.525838] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a27419e5-3e74-4314-9bbd-0b691abe13f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.586423] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350090, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.680564] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1050.680801] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1050.680991] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleting the datastore file [datastore2] e271159e-14ed-4a88-a7b8-23d404d77985 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.681353] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e19b36d1-0361-4c92-8a5f-a85dd59750e9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.690165] env[63355]: DEBUG oslo_vmware.api [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1050.690165] env[63355]: value = "task-1350092" [ 1050.690165] env[63355]: _type = "Task" [ 1050.690165] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.698536] env[63355]: DEBUG oslo_vmware.api [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350092, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.778568] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350088, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523534} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.778786] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 50e7499c-a8c1-41a6-adac-36e6c3b92819/50e7499c-a8c1-41a6-adac-36e6c3b92819.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1050.779196] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1050.779486] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6b3e880c-bc9f-4c79-8024-7e192ee7a60f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.782143] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4e8f4bc9-037a-4640-b8c1-71cfd9616c2e tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.702s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.788944] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1050.788944] env[63355]: value = "task-1350093" [ 1050.788944] env[63355]: _type = "Task" [ 1050.788944] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.796716] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350093, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.835357] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.059111] env[63355]: DEBUG nova.objects.instance [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'flavor' on Instance uuid 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.088099] env[63355]: DEBUG oslo_vmware.api [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350090, 'name': PowerOnVM_Task, 'duration_secs': 0.796028} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.088420] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1051.088658] env[63355]: INFO nova.compute.manager [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Took 8.23 seconds to spawn the instance on the hypervisor. [ 1051.088880] env[63355]: DEBUG nova.compute.manager [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.089709] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43eb36b-f593-4dd1-a927-cedca9063b93 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.201504] env[63355]: DEBUG oslo_vmware.api [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350092, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.403579} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.201858] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1051.202078] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1051.202266] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1051.202444] env[63355]: INFO nova.compute.manager [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1051.202690] env[63355]: DEBUG oslo.service.loopingcall [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1051.202891] env[63355]: DEBUG nova.compute.manager [-] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1051.202989] env[63355]: DEBUG nova.network.neutron [-] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1051.298793] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350093, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089141} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.299078] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1051.299841] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab6be9c-6a52-42d8-9044-8267fafb50f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.322286] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 50e7499c-a8c1-41a6-adac-36e6c3b92819/50e7499c-a8c1-41a6-adac-36e6c3b92819.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1051.322540] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59ac4af7-35d9-4617-a113-f68e8983f677 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.344432] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.345694] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1051.345694] env[63355]: value = "task-1350094" [ 1051.345694] env[63355]: _type = "Task" [ 1051.345694] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.353148] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350094, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.472314] env[63355]: DEBUG nova.compute.manager [req-956c025b-be15-40e7-a04f-adceb3125456 req-1d0119cd-8916-4e18-a53a-7b57113ae7ca service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Received event network-vif-deleted-93a72bba-8664-4305-a04f-e5e0449a27ff {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.472471] env[63355]: INFO nova.compute.manager [req-956c025b-be15-40e7-a04f-adceb3125456 req-1d0119cd-8916-4e18-a53a-7b57113ae7ca service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Neutron deleted interface 93a72bba-8664-4305-a04f-e5e0449a27ff; detaching it from the instance and deleting it from the info cache [ 1051.472780] env[63355]: DEBUG nova.network.neutron [req-956c025b-be15-40e7-a04f-adceb3125456 req-1d0119cd-8916-4e18-a53a-7b57113ae7ca service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.564788] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.564788] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.564788] env[63355]: DEBUG nova.network.neutron [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1051.564788] env[63355]: DEBUG nova.objects.instance [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'info_cache' on Instance uuid 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.605910] env[63355]: INFO nova.compute.manager [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Took 16.14 seconds to build instance. [ 1051.845541] env[63355]: DEBUG oslo_vmware.api [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350079, 'name': ReconfigVM_Task, 'duration_secs': 5.962203} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.845794] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.846014] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Reconfigured VM to detach interface {{(pid=63355) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1051.855614] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350094, 'name': ReconfigVM_Task, 'duration_secs': 0.327271} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.855693] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 50e7499c-a8c1-41a6-adac-36e6c3b92819/50e7499c-a8c1-41a6-adac-36e6c3b92819.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.856311] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3898da2-8845-4faf-a37e-e19bab595860 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.862907] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1051.862907] env[63355]: value = "task-1350095" [ 1051.862907] env[63355]: _type = "Task" [ 1051.862907] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.869357] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350095, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.945750] env[63355]: DEBUG nova.network.neutron [-] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.975628] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f12c1506-208f-4627-a460-46e8157c66f9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.984806] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed096a1-18a2-49c2-8745-3be18d358ab2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.019775] env[63355]: DEBUG nova.compute.manager [req-956c025b-be15-40e7-a04f-adceb3125456 req-1d0119cd-8916-4e18-a53a-7b57113ae7ca service nova] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Detach interface failed, port_id=93a72bba-8664-4305-a04f-e5e0449a27ff, reason: Instance e271159e-14ed-4a88-a7b8-23d404d77985 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1052.068456] env[63355]: DEBUG nova.objects.base [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Object Instance<537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5> lazy-loaded attributes: flavor,info_cache {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1052.107849] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b46d3c1d-0cfe-4c68-ab17-1e63a9b3345c tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "e398ee67-eb02-4256-9120-06111f110692" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.654s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.184228] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1587e3e2-cbf6-4023-a67c-3fa8c9477be2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.192879] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-892127b7-00b9-4ab3-bcee-30ee48eac5a2 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Suspending the VM {{(pid=63355) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1052.193176] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-adfb2180-8759-4be0-b945-4eca39696291 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.200599] env[63355]: DEBUG oslo_vmware.api [None req-892127b7-00b9-4ab3-bcee-30ee48eac5a2 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1052.200599] env[63355]: value = "task-1350096" [ 1052.200599] env[63355]: _type = "Task" [ 1052.200599] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.209315] env[63355]: DEBUG oslo_vmware.api [None req-892127b7-00b9-4ab3-bcee-30ee48eac5a2 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350096, 'name': SuspendVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.372486] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350095, 'name': Rename_Task, 'duration_secs': 0.16984} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.372816] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1052.373127] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b608807-cdea-4706-ab7d-bc3ec030df14 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.380601] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1052.380601] env[63355]: value = "task-1350097" [ 1052.380601] env[63355]: _type = "Task" [ 1052.380601] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.388329] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350097, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.449299] env[63355]: INFO nova.compute.manager [-] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Took 1.25 seconds to deallocate network for instance. [ 1052.713015] env[63355]: DEBUG oslo_vmware.api [None req-892127b7-00b9-4ab3-bcee-30ee48eac5a2 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350096, 'name': SuspendVM_Task} progress is 54%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.799124] env[63355]: DEBUG nova.network.neutron [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance_info_cache with network_info: [{"id": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "address": "fa:16:3e:b3:26:57", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap976ab0e0-5c", "ovs_interfaceid": "976ab0e0-5cea-4265-bc51-9b5cd29ed954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.890839] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350097, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.956633] env[63355]: DEBUG oslo_concurrency.lockutils [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.957029] env[63355]: DEBUG oslo_concurrency.lockutils [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.957308] env[63355]: DEBUG nova.objects.instance [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lazy-loading 'resources' on Instance uuid e271159e-14ed-4a88-a7b8-23d404d77985 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.192771] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.192978] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.193190] env[63355]: DEBUG nova.network.neutron [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1053.212778] env[63355]: DEBUG oslo_vmware.api [None req-892127b7-00b9-4ab3-bcee-30ee48eac5a2 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350096, 'name': SuspendVM_Task, 'duration_secs': 0.95771} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.213058] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-892127b7-00b9-4ab3-bcee-30ee48eac5a2 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Suspended the VM {{(pid=63355) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1053.213251] env[63355]: DEBUG nova.compute.manager [None req-892127b7-00b9-4ab3-bcee-30ee48eac5a2 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.214016] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d02649-8257-4754-bfbf-5891e8206594 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.302189] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.392962] env[63355]: DEBUG oslo_vmware.api [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350097, 'name': PowerOnVM_Task, 'duration_secs': 0.739334} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.393292] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1053.393416] env[63355]: INFO nova.compute.manager [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Took 8.23 seconds to spawn the instance on the hypervisor. [ 1053.393600] env[63355]: DEBUG nova.compute.manager [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.394380] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d473ba3-3cf0-4ffa-8d64-445d6d80feec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.575019] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed982f2f-a641-4120-a160-618632e3719b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.582743] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf93d5a-57c6-42a4-b84f-9c6d39c4ca16 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.611843] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcc9e07-5747-455b-8fa7-c508e5e95845 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.618859] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe1b1f3-727d-49c5-a313-7987fe3d9c84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.632024] env[63355]: DEBUG nova.compute.provider_tree [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.805737] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1053.807238] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67de8008-372f-4739-81a5-1e39295e8bfe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.813504] env[63355]: DEBUG oslo_vmware.api [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1053.813504] env[63355]: value = "task-1350098" [ 1053.813504] env[63355]: _type = "Task" [ 1053.813504] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.821649] env[63355]: DEBUG oslo_vmware.api [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350098, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.898007] env[63355]: INFO nova.network.neutron [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Port 563020c5-4b41-4820-9807-82cf864fcade from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1053.898480] env[63355]: DEBUG nova.network.neutron [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [{"id": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "address": "fa:16:3e:08:9e:2a", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5a40b6-5d", "ovs_interfaceid": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.902149] env[63355]: DEBUG nova.compute.manager [req-f45c214a-1115-4dcc-89c9-f2826e161c26 req-782517c5-55d6-46f0-8839-08071a82a2db service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received event network-changed-aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.902347] env[63355]: DEBUG nova.compute.manager [req-f45c214a-1115-4dcc-89c9-f2826e161c26 req-782517c5-55d6-46f0-8839-08071a82a2db service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing instance network info cache due to event network-changed-aa5a40b6-5d12-4d62-88d2-32a26a65a241. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1053.902554] env[63355]: DEBUG oslo_concurrency.lockutils [req-f45c214a-1115-4dcc-89c9-f2826e161c26 req-782517c5-55d6-46f0-8839-08071a82a2db service nova] Acquiring lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.910756] env[63355]: INFO nova.compute.manager [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Took 15.12 seconds to build instance. [ 1054.135354] env[63355]: DEBUG nova.scheduler.client.report [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.323523] env[63355]: DEBUG oslo_vmware.api [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350098, 'name': PowerOnVM_Task, 'duration_secs': 0.485219} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.323811] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1054.324042] env[63355]: DEBUG nova.compute.manager [None req-5f7d5682-df9e-4528-8d8b-fe8eb1804eaa tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1054.324889] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47d60f8-1f99-4671-b128-9c13e045e177 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.404318] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.406541] env[63355]: DEBUG oslo_concurrency.lockutils [req-f45c214a-1115-4dcc-89c9-f2826e161c26 req-782517c5-55d6-46f0-8839-08071a82a2db service nova] Acquired lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.406760] env[63355]: DEBUG nova.network.neutron [req-f45c214a-1115-4dcc-89c9-f2826e161c26 req-782517c5-55d6-46f0-8839-08071a82a2db service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Refreshing network info cache for port aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1054.412606] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b874c9db-7b23-46b1-a5ce-471cd464d82c tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.628s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.639273] env[63355]: DEBUG oslo_concurrency.lockutils [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.657093] env[63355]: INFO nova.scheduler.client.report [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted allocations for instance e271159e-14ed-4a88-a7b8-23d404d77985 [ 1054.672584] env[63355]: DEBUG oslo_concurrency.lockutils [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "interface-fa748748-31ec-431a-a628-5ea179e26fc8-563020c5-4b41-4820-9807-82cf864fcade" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.672947] env[63355]: DEBUG oslo_concurrency.lockutils [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-fa748748-31ec-431a-a628-5ea179e26fc8-563020c5-4b41-4820-9807-82cf864fcade" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.673465] env[63355]: DEBUG nova.objects.instance [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'flavor' on Instance uuid fa748748-31ec-431a-a628-5ea179e26fc8 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.718597] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "e398ee67-eb02-4256-9120-06111f110692" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.718872] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "e398ee67-eb02-4256-9120-06111f110692" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.719131] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "e398ee67-eb02-4256-9120-06111f110692-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.719386] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "e398ee67-eb02-4256-9120-06111f110692-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.719526] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "e398ee67-eb02-4256-9120-06111f110692-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.721514] env[63355]: INFO nova.compute.manager [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Terminating instance [ 1054.723297] env[63355]: DEBUG nova.compute.manager [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1054.723486] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1054.724325] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a275ab-b629-4640-9036-b891acfb5152 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.732347] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1054.732515] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-edf7000c-9fb1-437f-b9ed-1eeb93f3a11d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.799841] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1054.800123] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1054.800318] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleting the datastore file [datastore1] e398ee67-eb02-4256-9120-06111f110692 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.800668] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5cab2c86-df81-4e90-9e97-faaed36ec632 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.807053] env[63355]: DEBUG oslo_vmware.api [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1054.807053] env[63355]: value = "task-1350100" [ 1054.807053] env[63355]: _type = "Task" [ 1054.807053] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.814588] env[63355]: DEBUG oslo_vmware.api [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350100, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.909028] env[63355]: DEBUG oslo_concurrency.lockutils [None req-6694514b-0903-48d3-bde4-4c40db34cca6 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-da3cb83c-6368-49c0-9b11-0498221e3c0f-563020c5-4b41-4820-9807-82cf864fcade" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.662s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.143252] env[63355]: DEBUG nova.network.neutron [req-f45c214a-1115-4dcc-89c9-f2826e161c26 req-782517c5-55d6-46f0-8839-08071a82a2db service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updated VIF entry in instance network info cache for port aa5a40b6-5d12-4d62-88d2-32a26a65a241. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1055.143641] env[63355]: DEBUG nova.network.neutron [req-f45c214a-1115-4dcc-89c9-f2826e161c26 req-782517c5-55d6-46f0-8839-08071a82a2db service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [{"id": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "address": "fa:16:3e:08:9e:2a", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa5a40b6-5d", "ovs_interfaceid": "aa5a40b6-5d12-4d62-88d2-32a26a65a241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.165798] env[63355]: DEBUG oslo_concurrency.lockutils [None req-071d504f-539d-4d59-8631-e6714d91a6a5 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "e271159e-14ed-4a88-a7b8-23d404d77985" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.173s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.270844] env[63355]: DEBUG nova.objects.instance [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'pci_requests' on Instance uuid fa748748-31ec-431a-a628-5ea179e26fc8 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.316829] env[63355]: DEBUG oslo_vmware.api [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350100, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.337593} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.317173] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1055.317366] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1055.317546] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1055.317716] env[63355]: INFO nova.compute.manager [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: e398ee67-eb02-4256-9120-06111f110692] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1055.318143] env[63355]: DEBUG oslo.service.loopingcall [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.318380] env[63355]: DEBUG nova.compute.manager [-] [instance: e398ee67-eb02-4256-9120-06111f110692] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1055.318480] env[63355]: DEBUG nova.network.neutron [-] [instance: e398ee67-eb02-4256-9120-06111f110692] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1055.646162] env[63355]: DEBUG oslo_concurrency.lockutils [req-f45c214a-1115-4dcc-89c9-f2826e161c26 req-782517c5-55d6-46f0-8839-08071a82a2db service nova] Releasing lock "refresh_cache-da3cb83c-6368-49c0-9b11-0498221e3c0f" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.728669] env[63355]: DEBUG nova.compute.manager [req-927d59fa-fff4-466e-8a85-62254b67c47e req-520f1474-4242-4af6-884f-f98f977131c6 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Received event network-vif-deleted-6a28601f-1caa-4d3b-97c2-d913341d3bb9 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.728888] env[63355]: INFO nova.compute.manager [req-927d59fa-fff4-466e-8a85-62254b67c47e req-520f1474-4242-4af6-884f-f98f977131c6 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Neutron deleted interface 6a28601f-1caa-4d3b-97c2-d913341d3bb9; detaching it from the instance and deleting it from the info cache [ 1055.729178] env[63355]: DEBUG nova.network.neutron [req-927d59fa-fff4-466e-8a85-62254b67c47e req-520f1474-4242-4af6-884f-f98f977131c6 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.774255] env[63355]: DEBUG nova.objects.base [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1055.774488] env[63355]: DEBUG nova.network.neutron [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1055.792798] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.793126] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.793379] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.793602] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.793798] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.796282] env[63355]: INFO nova.compute.manager [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Terminating instance [ 1055.798830] env[63355]: DEBUG nova.compute.manager [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1055.799011] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1055.799861] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e8fb93-d73f-454f-9146-84fbec44a862 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.807712] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1055.808021] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0864eed8-8773-4423-9de7-2bc52cac7784 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.814486] env[63355]: DEBUG oslo_vmware.api [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1055.814486] env[63355]: value = "task-1350101" [ 1055.814486] env[63355]: _type = "Task" [ 1055.814486] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.821930] env[63355]: DEBUG oslo_vmware.api [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350101, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.841446] env[63355]: DEBUG nova.policy [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77243643ea724b72858a8682a2a054a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d47b671ea9c429391cbdae7e24adadf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1055.902139] env[63355]: DEBUG nova.compute.manager [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Stashing vm_state: active {{(pid=63355) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1055.932812] env[63355]: DEBUG nova.compute.manager [req-56afe80a-2315-49be-9cda-a41afbb43678 req-78f220b5-82b3-4d3f-b7e3-43677272597e service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Received event network-changed-546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.933491] env[63355]: DEBUG nova.compute.manager [req-56afe80a-2315-49be-9cda-a41afbb43678 req-78f220b5-82b3-4d3f-b7e3-43677272597e service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Refreshing instance network info cache due to event network-changed-546aac3d-2d71-48bd-84ac-6161e42a90dd. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1055.933544] env[63355]: DEBUG oslo_concurrency.lockutils [req-56afe80a-2315-49be-9cda-a41afbb43678 req-78f220b5-82b3-4d3f-b7e3-43677272597e service nova] Acquiring lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.933824] env[63355]: DEBUG oslo_concurrency.lockutils [req-56afe80a-2315-49be-9cda-a41afbb43678 req-78f220b5-82b3-4d3f-b7e3-43677272597e service nova] Acquired lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.934117] env[63355]: DEBUG nova.network.neutron [req-56afe80a-2315-49be-9cda-a41afbb43678 req-78f220b5-82b3-4d3f-b7e3-43677272597e service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Refreshing network info cache for port 546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1056.057471] env[63355]: DEBUG nova.network.neutron [-] [instance: e398ee67-eb02-4256-9120-06111f110692] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.231921] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18cbbea3-e144-4b5b-859e-fde8bdaf1c25 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.241385] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8924ad9e-f737-4690-ba66-ef8e542ab2eb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.267245] env[63355]: DEBUG nova.compute.manager [req-927d59fa-fff4-466e-8a85-62254b67c47e req-520f1474-4242-4af6-884f-f98f977131c6 service nova] [instance: e398ee67-eb02-4256-9120-06111f110692] Detach interface failed, port_id=6a28601f-1caa-4d3b-97c2-d913341d3bb9, reason: Instance e398ee67-eb02-4256-9120-06111f110692 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1056.271245] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.271464] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.324855] env[63355]: DEBUG oslo_vmware.api [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350101, 'name': PowerOffVM_Task, 'duration_secs': 0.170824} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.325105] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1056.325286] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1056.325548] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cda75748-f686-4c2a-bb81-a9d59e8b65c8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.387887] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1056.388132] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1056.388331] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleting the datastore file [datastore1] 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.388599] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-173d251f-a742-48c8-8083-26fedc436525 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.394973] env[63355]: DEBUG oslo_vmware.api [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1056.394973] env[63355]: value = "task-1350103" [ 1056.394973] env[63355]: _type = "Task" [ 1056.394973] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.402905] env[63355]: DEBUG oslo_vmware.api [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350103, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.422549] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.422837] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.563688] env[63355]: INFO nova.compute.manager [-] [instance: e398ee67-eb02-4256-9120-06111f110692] Took 1.25 seconds to deallocate network for instance. [ 1056.676382] env[63355]: DEBUG nova.network.neutron [req-56afe80a-2315-49be-9cda-a41afbb43678 req-78f220b5-82b3-4d3f-b7e3-43677272597e service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updated VIF entry in instance network info cache for port 546aac3d-2d71-48bd-84ac-6161e42a90dd. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1056.676760] env[63355]: DEBUG nova.network.neutron [req-56afe80a-2315-49be-9cda-a41afbb43678 req-78f220b5-82b3-4d3f-b7e3-43677272597e service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updating instance_info_cache with network_info: [{"id": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "address": "fa:16:3e:40:51:98", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546aac3d-2d", "ovs_interfaceid": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.774779] env[63355]: DEBUG nova.compute.manager [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1056.905042] env[63355]: DEBUG oslo_vmware.api [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350103, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.927627] env[63355]: INFO nova.compute.claims [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.069761] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.180246] env[63355]: DEBUG oslo_concurrency.lockutils [req-56afe80a-2315-49be-9cda-a41afbb43678 req-78f220b5-82b3-4d3f-b7e3-43677272597e service nova] Releasing lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.293350] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.333152] env[63355]: DEBUG nova.network.neutron [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Successfully updated port: 563020c5-4b41-4820-9807-82cf864fcade {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.406411] env[63355]: DEBUG oslo_vmware.api [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350103, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.686603} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.406745] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.406998] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1057.407211] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1057.407391] env[63355]: INFO nova.compute.manager [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1057.407628] env[63355]: DEBUG oslo.service.loopingcall [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.407816] env[63355]: DEBUG nova.compute.manager [-] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1057.407912] env[63355]: DEBUG nova.network.neutron [-] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1057.432892] env[63355]: INFO nova.compute.resource_tracker [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating resource usage from migration bc6e8ef9-18e4-43c9-a2db-1b629c809422 [ 1057.545344] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2952aca3-fa5b-4a72-905f-e239f4557595 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.553205] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2f12d8-70a8-4e3e-9089-041a0effca49 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.581974] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8175fe98-8b51-465f-b9c2-354b50f0fdc5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.589064] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488cdbd0-e0e2-438e-bc08-1b41f0b68a31 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.601780] env[63355]: DEBUG nova.compute.provider_tree [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.753535] env[63355]: DEBUG nova.compute.manager [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Received event network-vif-plugged-563020c5-4b41-4820-9807-82cf864fcade {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.753773] env[63355]: DEBUG oslo_concurrency.lockutils [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] Acquiring lock "fa748748-31ec-431a-a628-5ea179e26fc8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.753953] env[63355]: DEBUG oslo_concurrency.lockutils [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] Lock "fa748748-31ec-431a-a628-5ea179e26fc8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.754267] env[63355]: DEBUG oslo_concurrency.lockutils [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] Lock "fa748748-31ec-431a-a628-5ea179e26fc8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.754511] env[63355]: DEBUG nova.compute.manager [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] No waiting events found dispatching network-vif-plugged-563020c5-4b41-4820-9807-82cf864fcade {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1057.754621] env[63355]: WARNING nova.compute.manager [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Received unexpected event network-vif-plugged-563020c5-4b41-4820-9807-82cf864fcade for instance with vm_state active and task_state None. [ 1057.754785] env[63355]: DEBUG nova.compute.manager [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Received event network-changed-563020c5-4b41-4820-9807-82cf864fcade {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.755018] env[63355]: DEBUG nova.compute.manager [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Refreshing instance network info cache due to event network-changed-563020c5-4b41-4820-9807-82cf864fcade. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1057.755226] env[63355]: DEBUG oslo_concurrency.lockutils [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] Acquiring lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.755367] env[63355]: DEBUG oslo_concurrency.lockutils [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] Acquired lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.755524] env[63355]: DEBUG nova.network.neutron [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Refreshing network info cache for port 563020c5-4b41-4820-9807-82cf864fcade {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1057.836116] env[63355]: DEBUG oslo_concurrency.lockutils [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.105276] env[63355]: DEBUG nova.scheduler.client.report [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.316265] env[63355]: DEBUG nova.network.neutron [-] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.472585] env[63355]: DEBUG nova.network.neutron [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Added VIF to instance network info cache for port 563020c5-4b41-4820-9807-82cf864fcade. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 1058.473016] env[63355]: DEBUG nova.network.neutron [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updating instance_info_cache with network_info: [{"id": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "address": "fa:16:3e:40:51:98", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546aac3d-2d", "ovs_interfaceid": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "563020c5-4b41-4820-9807-82cf864fcade", "address": "fa:16:3e:de:c3:bd", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap563020c5-4b", "ovs_interfaceid": "563020c5-4b41-4820-9807-82cf864fcade", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.610340] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.187s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.610608] env[63355]: INFO nova.compute.manager [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Migrating [ 1058.617220] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.548s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.617492] env[63355]: DEBUG nova.objects.instance [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lazy-loading 'resources' on Instance uuid e398ee67-eb02-4256-9120-06111f110692 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.819358] env[63355]: INFO nova.compute.manager [-] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Took 1.41 seconds to deallocate network for instance. [ 1058.975646] env[63355]: DEBUG oslo_concurrency.lockutils [req-533c7ff6-a91b-41b6-8a80-d5143bfa7fec req-2a29870e-3780-4fc7-8448-2812022953e8 service nova] Releasing lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.976195] env[63355]: DEBUG oslo_concurrency.lockutils [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.976400] env[63355]: DEBUG nova.network.neutron [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1059.127209] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.127394] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.127572] env[63355]: DEBUG nova.network.neutron [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1059.229054] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b749cd18-1f1f-435d-8d65-ef882042a0f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.236451] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c47bad-67d2-4c34-b657-717723533333 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.267139] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a8c0cac-784c-4dcd-b29e-a561c4f61ed8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.274693] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5362845b-86f1-4457-8af8-559dc2ab1605 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.931284] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.936190] env[63355]: DEBUG nova.compute.manager [req-f1831aa9-b798-494f-8ea5-4fc873413416 req-52f1de03-5ba7-4eb1-9619-bc19bb07c925 service nova] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Received event network-vif-deleted-976ab0e0-5cea-4265-bc51-9b5cd29ed954 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.948059] env[63355]: DEBUG nova.compute.provider_tree [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.982979] env[63355]: WARNING nova.network.neutron [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] 943d3ecf-0c77-4c51-a997-b17d86259c34 already exists in list: networks containing: ['943d3ecf-0c77-4c51-a997-b17d86259c34']. ignoring it [ 1059.983220] env[63355]: WARNING nova.network.neutron [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] 943d3ecf-0c77-4c51-a997-b17d86259c34 already exists in list: networks containing: ['943d3ecf-0c77-4c51-a997-b17d86259c34']. ignoring it [ 1059.983422] env[63355]: WARNING nova.network.neutron [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] 563020c5-4b41-4820-9807-82cf864fcade already exists in list: port_ids containing: ['563020c5-4b41-4820-9807-82cf864fcade']. ignoring it [ 1060.260663] env[63355]: DEBUG nova.network.neutron [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance_info_cache with network_info: [{"id": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "address": "fa:16:3e:8a:f8:58", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba0aaec1-3b", "ovs_interfaceid": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.348279] env[63355]: DEBUG nova.network.neutron [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updating instance_info_cache with network_info: [{"id": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "address": "fa:16:3e:40:51:98", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546aac3d-2d", "ovs_interfaceid": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "563020c5-4b41-4820-9807-82cf864fcade", "address": "fa:16:3e:de:c3:bd", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap563020c5-4b", "ovs_interfaceid": "563020c5-4b41-4820-9807-82cf864fcade", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.451551] env[63355]: DEBUG nova.scheduler.client.report [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1060.765717] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.850449] env[63355]: DEBUG oslo_concurrency.lockutils [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.851234] env[63355]: DEBUG oslo_concurrency.lockutils [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.851439] env[63355]: DEBUG oslo_concurrency.lockutils [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.852327] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e09c646-c26e-4b14-8862-c87a3785178d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.868992] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1060.869295] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1060.869486] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1060.869681] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1060.869832] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1060.869983] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1060.870208] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1060.870369] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1060.870534] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1060.870696] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1060.870868] env[63355]: DEBUG nova.virt.hardware [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.877114] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Reconfiguring VM to attach interface {{(pid=63355) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1060.877485] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ed38283-e407-4706-a631-39a06a1459b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.893943] env[63355]: DEBUG oslo_vmware.api [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1060.893943] env[63355]: value = "task-1350104" [ 1060.893943] env[63355]: _type = "Task" [ 1060.893943] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.901848] env[63355]: DEBUG oslo_vmware.api [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350104, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.956137] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.339s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.959832] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.666s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.962309] env[63355]: INFO nova.compute.claims [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1060.981662] env[63355]: INFO nova.scheduler.client.report [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted allocations for instance e398ee67-eb02-4256-9120-06111f110692 [ 1061.404198] env[63355]: DEBUG oslo_vmware.api [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350104, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.493097] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3369e2b6-df84-4065-af7b-ca986d9c19db tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "e398ee67-eb02-4256-9120-06111f110692" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.774s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.903890] env[63355]: DEBUG oslo_vmware.api [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350104, 'name': ReconfigVM_Task, 'duration_secs': 0.707897} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.904419] env[63355]: DEBUG oslo_concurrency.lockutils [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.904646] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Reconfigured VM to attach interface {{(pid=63355) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1062.088767] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642f0ecb-75db-4bea-bbad-6335e7b1289d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.096500] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f625d24-b32d-4cf6-bea0-42e8c15c198a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.126222] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4874e7-ffcc-4779-868b-65fa50773d54 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.133826] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf911a2b-acce-4902-83fa-9690b1e1063f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.148967] env[63355]: DEBUG nova.compute.provider_tree [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.283627] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575bec56-9879-40eb-a0a3-60ef293d2c76 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.302521] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance '50e7499c-a8c1-41a6-adac-36e6c3b92819' progress to 0 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1062.409315] env[63355]: DEBUG oslo_concurrency.lockutils [None req-254ee504-2614-4e78-a6b5-d02f8ce0d451 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-fa748748-31ec-431a-a628-5ea179e26fc8-563020c5-4b41-4820-9807-82cf864fcade" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.736s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.651943] env[63355]: DEBUG nova.scheduler.client.report [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1062.808346] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1062.808671] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55ff54b7-1b6a-4751-87cc-31954cddac9e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.815846] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1062.815846] env[63355]: value = "task-1350105" [ 1062.815846] env[63355]: _type = "Task" [ 1062.815846] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.824064] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.026493] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "8fc3ea4e-d525-4583-b237-f895545fd3d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.026739] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.157417] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.198s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.157968] env[63355]: DEBUG nova.compute.manager [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1063.161264] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.230s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.161477] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.180377] env[63355]: INFO nova.scheduler.client.report [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleted allocations for instance 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5 [ 1063.325660] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.529027] env[63355]: DEBUG nova.compute.manager [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1063.663527] env[63355]: DEBUG nova.compute.utils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1063.665050] env[63355]: DEBUG nova.compute.manager [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1063.665236] env[63355]: DEBUG nova.network.neutron [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1063.686701] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a44cc737-91f4-4f2f-9e56-963409966781 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.893s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.703651] env[63355]: DEBUG nova.policy [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3721c4a73df54ae9b44110cb9e8590a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc547065748241e8ac7b6c499ddaea66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1063.825797] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350105, 'name': PowerOffVM_Task, 'duration_secs': 0.824266} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.826553] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1063.826553] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance '50e7499c-a8c1-41a6-adac-36e6c3b92819' progress to 17 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1063.854560] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "interface-fa748748-31ec-431a-a628-5ea179e26fc8-563020c5-4b41-4820-9807-82cf864fcade" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.854560] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-fa748748-31ec-431a-a628-5ea179e26fc8-563020c5-4b41-4820-9807-82cf864fcade" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.941980] env[63355]: DEBUG nova.network.neutron [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Successfully created port: 9b26dcfe-3e4b-4d6c-bd1f-c628bd055377 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1064.051680] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.051941] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.053468] env[63355]: INFO nova.compute.claims [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1064.168900] env[63355]: DEBUG nova.compute.manager [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1064.333408] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1064.333671] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1064.334056] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.334274] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1064.334626] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.334794] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1064.335121] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1064.335359] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1064.335543] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1064.335751] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1064.335882] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.342160] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-820d7208-baeb-49f9-993c-934032f2fbae {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.358517] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.358695] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.360533] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c367d23c-4389-4292-b937-0710d9574600 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.363145] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1064.363145] env[63355]: value = "task-1350106" [ 1064.363145] env[63355]: _type = "Task" [ 1064.363145] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.380282] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5af2433-3867-4971-a409-1b40dd0f16b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.386984] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350106, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.410021] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Reconfiguring VM to detach interface {{(pid=63355) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1064.410658] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50441e2f-4643-42c2-b5d9-ef34d9ed22eb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.429256] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1064.429256] env[63355]: value = "task-1350107" [ 1064.429256] env[63355]: _type = "Task" [ 1064.429256] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.437692] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.873115] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350106, 'name': ReconfigVM_Task, 'duration_secs': 0.139669} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.873452] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance '50e7499c-a8c1-41a6-adac-36e6c3b92819' progress to 33 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1064.938752] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.040157] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.040352] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.167487] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d453af0e-8fc1-4870-8bb5-fe69b1bc01dd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.175569] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1efa78b-6215-4df9-8797-6e2b599be921 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.179396] env[63355]: DEBUG nova.compute.manager [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1065.210665] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55af8c85-2452-40e3-b6fd-760332f1afaa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.219170] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1065.219443] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1065.219616] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1065.219805] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1065.219955] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1065.220120] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1065.220333] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1065.220496] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1065.220665] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1065.220829] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1065.221011] env[63355]: DEBUG nova.virt.hardware [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1065.222222] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-321dc264-0ca4-4579-83c5-8a005fcd2bf7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.226475] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03503f0a-5b13-4e10-b3da-b94067f76deb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.238509] env[63355]: DEBUG nova.compute.provider_tree [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.242509] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb103f63-f1bb-4576-87fc-67c576f20993 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.379528] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1065.379739] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1065.379903] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1065.380104] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1065.380260] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1065.380470] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1065.380718] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1065.380887] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1065.381072] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1065.381286] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1065.381473] env[63355]: DEBUG nova.virt.hardware [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1065.386953] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1065.387600] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8996246e-5be6-47d4-b42b-d95936e8b559 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.406501] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1065.406501] env[63355]: value = "task-1350108" [ 1065.406501] env[63355]: _type = "Task" [ 1065.406501] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.415039] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350108, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.439121] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.507592] env[63355]: DEBUG nova.compute.manager [req-ff265970-0c16-4cb9-8a2c-b201c00dd69a req-1b87aa07-38de-4cbd-b3d6-2aafc50813d3 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Received event network-vif-plugged-9b26dcfe-3e4b-4d6c-bd1f-c628bd055377 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.507840] env[63355]: DEBUG oslo_concurrency.lockutils [req-ff265970-0c16-4cb9-8a2c-b201c00dd69a req-1b87aa07-38de-4cbd-b3d6-2aafc50813d3 service nova] Acquiring lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.508520] env[63355]: DEBUG oslo_concurrency.lockutils [req-ff265970-0c16-4cb9-8a2c-b201c00dd69a req-1b87aa07-38de-4cbd-b3d6-2aafc50813d3 service nova] Lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.508520] env[63355]: DEBUG oslo_concurrency.lockutils [req-ff265970-0c16-4cb9-8a2c-b201c00dd69a req-1b87aa07-38de-4cbd-b3d6-2aafc50813d3 service nova] Lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.508719] env[63355]: DEBUG nova.compute.manager [req-ff265970-0c16-4cb9-8a2c-b201c00dd69a req-1b87aa07-38de-4cbd-b3d6-2aafc50813d3 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] No waiting events found dispatching network-vif-plugged-9b26dcfe-3e4b-4d6c-bd1f-c628bd055377 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1065.508910] env[63355]: WARNING nova.compute.manager [req-ff265970-0c16-4cb9-8a2c-b201c00dd69a req-1b87aa07-38de-4cbd-b3d6-2aafc50813d3 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Received unexpected event network-vif-plugged-9b26dcfe-3e4b-4d6c-bd1f-c628bd055377 for instance with vm_state building and task_state spawning. [ 1065.542950] env[63355]: DEBUG nova.compute.manager [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1065.595540] env[63355]: DEBUG nova.network.neutron [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Successfully updated port: 9b26dcfe-3e4b-4d6c-bd1f-c628bd055377 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1065.747876] env[63355]: DEBUG nova.scheduler.client.report [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.916268] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350108, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.941355] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.064065] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.098955] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "refresh_cache-f08bca27-ab50-40d0-a6d9-4842a37a4faf" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.099114] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "refresh_cache-f08bca27-ab50-40d0-a6d9-4842a37a4faf" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.100044] env[63355]: DEBUG nova.network.neutron [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.253693] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.201s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.254200] env[63355]: DEBUG nova.compute.manager [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1066.256862] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.193s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.258755] env[63355]: INFO nova.compute.claims [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1066.416918] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350108, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.440287] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.631546] env[63355]: DEBUG nova.network.neutron [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1066.753125] env[63355]: DEBUG nova.network.neutron [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Updating instance_info_cache with network_info: [{"id": "9b26dcfe-3e4b-4d6c-bd1f-c628bd055377", "address": "fa:16:3e:24:1b:ea", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b26dcfe-3e", "ovs_interfaceid": "9b26dcfe-3e4b-4d6c-bd1f-c628bd055377", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.762179] env[63355]: DEBUG nova.compute.utils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1066.765379] env[63355]: DEBUG nova.compute.manager [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1066.765559] env[63355]: DEBUG nova.network.neutron [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1066.801822] env[63355]: DEBUG nova.policy [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'beb41983e69a44b2a108f552da5ef640', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfff1b9903264e5586119ebd3a3602de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1066.917801] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350108, 'name': ReconfigVM_Task, 'duration_secs': 1.191615} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.917801] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1066.918391] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d316a10e-2ec8-42c1-90b0-da877f09fbb5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.940963] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 50e7499c-a8c1-41a6-adac-36e6c3b92819/50e7499c-a8c1-41a6-adac-36e6c3b92819.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.943943] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e04a82d8-10a5-4697-8f56-703a2943ca85 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.962199] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.963400] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1066.963400] env[63355]: value = "task-1350109" [ 1066.963400] env[63355]: _type = "Task" [ 1066.963400] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.970697] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350109, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.042492] env[63355]: DEBUG nova.network.neutron [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Successfully created port: 6ca66441-d91f-4644-b218-fc67cf0b1ef8 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1067.256230] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "refresh_cache-f08bca27-ab50-40d0-a6d9-4842a37a4faf" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.256658] env[63355]: DEBUG nova.compute.manager [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Instance network_info: |[{"id": "9b26dcfe-3e4b-4d6c-bd1f-c628bd055377", "address": "fa:16:3e:24:1b:ea", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b26dcfe-3e", "ovs_interfaceid": "9b26dcfe-3e4b-4d6c-bd1f-c628bd055377", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1067.257131] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:1b:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b26dcfe-3e4b-4d6c-bd1f-c628bd055377', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1067.265572] env[63355]: DEBUG oslo.service.loopingcall [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.265817] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1067.266054] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-042a4241-1755-46af-aa01-7f2e97fd6d5d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.281844] env[63355]: DEBUG nova.compute.manager [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1067.293262] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1067.293262] env[63355]: value = "task-1350110" [ 1067.293262] env[63355]: _type = "Task" [ 1067.293262] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.303236] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350110, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.406816] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee5f024-8962-4158-a106-54e1fe18ae50 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.414018] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199d25bf-ef51-4696-a1fd-b4e00888bf61 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.445230] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69dbf03c-9af4-4837-90cc-1a88e2207742 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.454788] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a85444e-8b88-4439-a716-1f8fb4906141 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.458555] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.468063] env[63355]: DEBUG nova.compute.provider_tree [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.477765] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350109, 'name': ReconfigVM_Task, 'duration_secs': 0.271841} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.478790] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 50e7499c-a8c1-41a6-adac-36e6c3b92819/50e7499c-a8c1-41a6-adac-36e6c3b92819.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.479091] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance '50e7499c-a8c1-41a6-adac-36e6c3b92819' progress to 50 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1067.513540] env[63355]: DEBUG oslo_concurrency.lockutils [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.513540] env[63355]: DEBUG oslo_concurrency.lockutils [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.535247] env[63355]: DEBUG nova.compute.manager [req-7e361d15-c8df-4b43-b8dc-f29079471c56 req-a50c02da-bd58-4bf7-a733-54ff39dbc2d2 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Received event network-changed-9b26dcfe-3e4b-4d6c-bd1f-c628bd055377 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.535500] env[63355]: DEBUG nova.compute.manager [req-7e361d15-c8df-4b43-b8dc-f29079471c56 req-a50c02da-bd58-4bf7-a733-54ff39dbc2d2 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Refreshing instance network info cache due to event network-changed-9b26dcfe-3e4b-4d6c-bd1f-c628bd055377. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1067.535660] env[63355]: DEBUG oslo_concurrency.lockutils [req-7e361d15-c8df-4b43-b8dc-f29079471c56 req-a50c02da-bd58-4bf7-a733-54ff39dbc2d2 service nova] Acquiring lock "refresh_cache-f08bca27-ab50-40d0-a6d9-4842a37a4faf" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.535806] env[63355]: DEBUG oslo_concurrency.lockutils [req-7e361d15-c8df-4b43-b8dc-f29079471c56 req-a50c02da-bd58-4bf7-a733-54ff39dbc2d2 service nova] Acquired lock "refresh_cache-f08bca27-ab50-40d0-a6d9-4842a37a4faf" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.535968] env[63355]: DEBUG nova.network.neutron [req-7e361d15-c8df-4b43-b8dc-f29079471c56 req-a50c02da-bd58-4bf7-a733-54ff39dbc2d2 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Refreshing network info cache for port 9b26dcfe-3e4b-4d6c-bd1f-c628bd055377 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.804032] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350110, 'name': CreateVM_Task, 'duration_secs': 0.2736} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.804032] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1067.804759] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.804950] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.805297] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1067.805557] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5592b5dd-77ff-4e09-a2ff-642811f589ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.809802] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1067.809802] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5207ea64-0059-e956-d67d-7097a6cbed15" [ 1067.809802] env[63355]: _type = "Task" [ 1067.809802] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.816875] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5207ea64-0059-e956-d67d-7097a6cbed15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.952769] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.972041] env[63355]: DEBUG nova.scheduler.client.report [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.987073] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837e9708-098e-4ca9-8a33-2a5d7fa13eaf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.006128] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0d9c4b-d80d-4e74-a81e-e3befaa33f33 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.023668] env[63355]: INFO nova.compute.manager [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Detaching volume 64055852-0585-4516-8f95-0996e47ccc04 [ 1068.025676] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance '50e7499c-a8c1-41a6-adac-36e6c3b92819' progress to 67 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1068.066948] env[63355]: INFO nova.virt.block_device [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Attempting to driver detach volume 64055852-0585-4516-8f95-0996e47ccc04 from mountpoint /dev/sdb [ 1068.067296] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1068.067616] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287741', 'volume_id': '64055852-0585-4516-8f95-0996e47ccc04', 'name': 'volume-64055852-0585-4516-8f95-0996e47ccc04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae3961d2-dc5b-4e49-acca-6fb52291f23b', 'attached_at': '', 'detached_at': '', 'volume_id': '64055852-0585-4516-8f95-0996e47ccc04', 'serial': '64055852-0585-4516-8f95-0996e47ccc04'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1068.068752] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3551a0d2-5732-4221-83eb-4c87701d852d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.092769] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769f6375-4a09-4d46-a946-4f7e73b44339 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.100099] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2707bb2b-b4e4-45bd-96e5-c45bec8fcfde {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.124725] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4992a9c-0853-4a02-894f-b156b9884144 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.139366] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] The volume has not been displaced from its original location: [datastore2] volume-64055852-0585-4516-8f95-0996e47ccc04/volume-64055852-0585-4516-8f95-0996e47ccc04.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1068.144465] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Reconfiguring VM instance instance-00000053 to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1068.146571] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8291104e-dc83-4c02-9b4f-5ee26a227080 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.164881] env[63355]: DEBUG oslo_vmware.api [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1068.164881] env[63355]: value = "task-1350111" [ 1068.164881] env[63355]: _type = "Task" [ 1068.164881] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.172581] env[63355]: DEBUG oslo_vmware.api [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350111, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.271831] env[63355]: DEBUG nova.network.neutron [req-7e361d15-c8df-4b43-b8dc-f29079471c56 req-a50c02da-bd58-4bf7-a733-54ff39dbc2d2 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Updated VIF entry in instance network info cache for port 9b26dcfe-3e4b-4d6c-bd1f-c628bd055377. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1068.272243] env[63355]: DEBUG nova.network.neutron [req-7e361d15-c8df-4b43-b8dc-f29079471c56 req-a50c02da-bd58-4bf7-a733-54ff39dbc2d2 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Updating instance_info_cache with network_info: [{"id": "9b26dcfe-3e4b-4d6c-bd1f-c628bd055377", "address": "fa:16:3e:24:1b:ea", "network": {"id": "064b22dc-e735-4f2a-9ef0-1c0310de87ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-2144636678-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc547065748241e8ac7b6c499ddaea66", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b26dcfe-3e", "ovs_interfaceid": "9b26dcfe-3e4b-4d6c-bd1f-c628bd055377", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.290742] env[63355]: DEBUG nova.compute.manager [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1068.316431] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1068.316702] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1068.316865] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1068.317199] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1068.317287] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1068.317425] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1068.317639] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1068.317802] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1068.317973] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1068.318172] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1068.318397] env[63355]: DEBUG nova.virt.hardware [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1068.319607] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742f8956-35e1-4c71-bdcf-29a1014e8fe5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.325993] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5207ea64-0059-e956-d67d-7097a6cbed15, 'name': SearchDatastore_Task, 'duration_secs': 0.0091} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.326783] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.327062] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1068.327351] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.327515] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.327737] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1068.328086] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-342d243b-8125-4d93-9984-c0ffe0380187 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.333148] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01138bdb-6ed5-4f64-b9fa-ab03259d29f4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.349230] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1068.349426] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1068.350128] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f046a4b7-47d1-4662-b266-2d2b65b69c11 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.356159] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1068.356159] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52139579-b52d-7067-5112-e439352e4a25" [ 1068.356159] env[63355]: _type = "Task" [ 1068.356159] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.363465] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52139579-b52d-7067-5112-e439352e4a25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.452388] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.477349] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.220s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.478064] env[63355]: DEBUG nova.compute.manager [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1068.505752] env[63355]: DEBUG nova.network.neutron [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Successfully updated port: 6ca66441-d91f-4644-b218-fc67cf0b1ef8 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1068.581783] env[63355]: DEBUG nova.network.neutron [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Port ba0aaec1-3b3e-4594-bf65-aa18843e439d binding to destination host cpu-1 is already ACTIVE {{(pid=63355) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1068.676014] env[63355]: DEBUG oslo_vmware.api [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350111, 'name': ReconfigVM_Task, 'duration_secs': 0.213137} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.676300] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Reconfigured VM instance instance-00000053 to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1068.680857] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1543491a-65bf-4b39-ac32-a990dae3eb2f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.695558] env[63355]: DEBUG oslo_vmware.api [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1068.695558] env[63355]: value = "task-1350112" [ 1068.695558] env[63355]: _type = "Task" [ 1068.695558] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.703224] env[63355]: DEBUG oslo_vmware.api [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350112, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.774628] env[63355]: DEBUG oslo_concurrency.lockutils [req-7e361d15-c8df-4b43-b8dc-f29079471c56 req-a50c02da-bd58-4bf7-a733-54ff39dbc2d2 service nova] Releasing lock "refresh_cache-f08bca27-ab50-40d0-a6d9-4842a37a4faf" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.866463] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52139579-b52d-7067-5112-e439352e4a25, 'name': SearchDatastore_Task, 'duration_secs': 0.007951} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.867122] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ba9a80a-9696-414c-a885-481a0b1c205e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.872428] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1068.872428] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52be4bdb-3a97-f619-fad1-e9ea8ff3d0cb" [ 1068.872428] env[63355]: _type = "Task" [ 1068.872428] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.879608] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52be4bdb-3a97-f619-fad1-e9ea8ff3d0cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.952281] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.982839] env[63355]: DEBUG nova.compute.utils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1068.984465] env[63355]: DEBUG nova.compute.manager [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1068.984648] env[63355]: DEBUG nova.network.neutron [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1069.008013] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.008197] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.008323] env[63355]: DEBUG nova.network.neutron [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1069.022913] env[63355]: DEBUG nova.policy [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54ad2682257b4250a8a96f4ba6a9523b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b3fee9bc99d49ea9de53d5dce52c79d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1069.206239] env[63355]: DEBUG oslo_vmware.api [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350112, 'name': ReconfigVM_Task, 'duration_secs': 0.136624} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.206797] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287741', 'volume_id': '64055852-0585-4516-8f95-0996e47ccc04', 'name': 'volume-64055852-0585-4516-8f95-0996e47ccc04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae3961d2-dc5b-4e49-acca-6fb52291f23b', 'attached_at': '', 'detached_at': '', 'volume_id': '64055852-0585-4516-8f95-0996e47ccc04', 'serial': '64055852-0585-4516-8f95-0996e47ccc04'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1069.280916] env[63355]: DEBUG nova.network.neutron [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Successfully created port: 2689b07f-15d6-4cad-b652-24c86298c985 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1069.383059] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52be4bdb-3a97-f619-fad1-e9ea8ff3d0cb, 'name': SearchDatastore_Task, 'duration_secs': 0.008858} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.383351] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.383615] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] f08bca27-ab50-40d0-a6d9-4842a37a4faf/f08bca27-ab50-40d0-a6d9-4842a37a4faf.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1069.383897] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e11c25b-9cf6-4393-a4cc-94187e0d1820 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.392303] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1069.392303] env[63355]: value = "task-1350113" [ 1069.392303] env[63355]: _type = "Task" [ 1069.392303] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.400162] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350113, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.453804] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.487712] env[63355]: DEBUG nova.compute.manager [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1069.544481] env[63355]: DEBUG nova.network.neutron [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1069.563094] env[63355]: DEBUG nova.compute.manager [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Received event network-vif-plugged-6ca66441-d91f-4644-b218-fc67cf0b1ef8 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.563391] env[63355]: DEBUG oslo_concurrency.lockutils [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] Acquiring lock "8fc3ea4e-d525-4583-b237-f895545fd3d1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.563577] env[63355]: DEBUG oslo_concurrency.lockutils [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.563724] env[63355]: DEBUG oslo_concurrency.lockutils [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.564112] env[63355]: DEBUG nova.compute.manager [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] No waiting events found dispatching network-vif-plugged-6ca66441-d91f-4644-b218-fc67cf0b1ef8 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1069.564317] env[63355]: WARNING nova.compute.manager [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Received unexpected event network-vif-plugged-6ca66441-d91f-4644-b218-fc67cf0b1ef8 for instance with vm_state building and task_state spawning. [ 1069.564489] env[63355]: DEBUG nova.compute.manager [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Received event network-changed-6ca66441-d91f-4644-b218-fc67cf0b1ef8 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.564667] env[63355]: DEBUG nova.compute.manager [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Refreshing instance network info cache due to event network-changed-6ca66441-d91f-4644-b218-fc67cf0b1ef8. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1069.564856] env[63355]: DEBUG oslo_concurrency.lockutils [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] Acquiring lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.606625] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "50e7499c-a8c1-41a6-adac-36e6c3b92819-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.606625] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.606965] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.717162] env[63355]: DEBUG nova.network.neutron [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance_info_cache with network_info: [{"id": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "address": "fa:16:3e:01:27:93", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca66441-d9", "ovs_interfaceid": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.755305] env[63355]: DEBUG nova.objects.instance [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'flavor' on Instance uuid ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.904139] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350113, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49917} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.904378] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] f08bca27-ab50-40d0-a6d9-4842a37a4faf/f08bca27-ab50-40d0-a6d9-4842a37a4faf.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1069.904708] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1069.904873] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1310384d-edbe-4ecd-9222-2341778bb96a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.911055] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1069.911055] env[63355]: value = "task-1350114" [ 1069.911055] env[63355]: _type = "Task" [ 1069.911055] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.918715] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350114, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.954832] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.220253] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.220410] env[63355]: DEBUG nova.compute.manager [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Instance network_info: |[{"id": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "address": "fa:16:3e:01:27:93", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca66441-d9", "ovs_interfaceid": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1070.220692] env[63355]: DEBUG oslo_concurrency.lockutils [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] Acquired lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.220886] env[63355]: DEBUG nova.network.neutron [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Refreshing network info cache for port 6ca66441-d91f-4644-b218-fc67cf0b1ef8 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1070.222115] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:27:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '52358fcc-0d9f-45dd-8c75-db533fd992c3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ca66441-d91f-4644-b218-fc67cf0b1ef8', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1070.229455] env[63355]: DEBUG oslo.service.loopingcall [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.232338] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1070.232729] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ce36065-6792-4f08-9fd8-8794d983f213 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.252645] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1070.252645] env[63355]: value = "task-1350115" [ 1070.252645] env[63355]: _type = "Task" [ 1070.252645] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.261994] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350115, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.420835] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350114, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059666} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.421158] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1070.421915] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a9fe8f-a0b8-4666-a099-290069059efe {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.443023] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] f08bca27-ab50-40d0-a6d9-4842a37a4faf/f08bca27-ab50-40d0-a6d9-4842a37a4faf.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.443888] env[63355]: DEBUG nova.network.neutron [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updated VIF entry in instance network info cache for port 6ca66441-d91f-4644-b218-fc67cf0b1ef8. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1070.444240] env[63355]: DEBUG nova.network.neutron [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance_info_cache with network_info: [{"id": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "address": "fa:16:3e:01:27:93", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca66441-d9", "ovs_interfaceid": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.445385] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf0c985e-9507-4cac-8792-32ab0df06da0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.470179] env[63355]: DEBUG oslo_vmware.api [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350107, 'name': ReconfigVM_Task, 'duration_secs': 5.782056} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.471377] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.471647] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Reconfigured VM to detach interface {{(pid=63355) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1070.473614] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1070.473614] env[63355]: value = "task-1350116" [ 1070.473614] env[63355]: _type = "Task" [ 1070.473614] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.482633] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350116, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.498630] env[63355]: DEBUG nova.compute.manager [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1070.526886] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1070.527232] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1070.527411] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.527607] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1070.527759] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.527911] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1070.528139] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1070.528304] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1070.528472] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1070.528637] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1070.528811] env[63355]: DEBUG nova.virt.hardware [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1070.529678] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ec8b38-a311-467a-bb55-e2b6515c7a7c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.538555] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a909e26b-95f3-4b44-b5c3-8b1b9c21fd94 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.648779] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.649069] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.649170] env[63355]: DEBUG nova.network.neutron [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1070.686029] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.717978] env[63355]: DEBUG nova.compute.manager [req-28e657f2-bd14-4e1c-8234-0ebfeb28b227 req-929b818d-f839-468a-8fb1-a199b3124bf8 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Received event network-vif-plugged-2689b07f-15d6-4cad-b652-24c86298c985 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1070.717978] env[63355]: DEBUG oslo_concurrency.lockutils [req-28e657f2-bd14-4e1c-8234-0ebfeb28b227 req-929b818d-f839-468a-8fb1-a199b3124bf8 service nova] Acquiring lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.717978] env[63355]: DEBUG oslo_concurrency.lockutils [req-28e657f2-bd14-4e1c-8234-0ebfeb28b227 req-929b818d-f839-468a-8fb1-a199b3124bf8 service nova] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.717978] env[63355]: DEBUG oslo_concurrency.lockutils [req-28e657f2-bd14-4e1c-8234-0ebfeb28b227 req-929b818d-f839-468a-8fb1-a199b3124bf8 service nova] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.717978] env[63355]: DEBUG nova.compute.manager [req-28e657f2-bd14-4e1c-8234-0ebfeb28b227 req-929b818d-f839-468a-8fb1-a199b3124bf8 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] No waiting events found dispatching network-vif-plugged-2689b07f-15d6-4cad-b652-24c86298c985 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1070.717978] env[63355]: WARNING nova.compute.manager [req-28e657f2-bd14-4e1c-8234-0ebfeb28b227 req-929b818d-f839-468a-8fb1-a199b3124bf8 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Received unexpected event network-vif-plugged-2689b07f-15d6-4cad-b652-24c86298c985 for instance with vm_state building and task_state spawning. [ 1070.763248] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350115, 'name': CreateVM_Task, 'duration_secs': 0.38572} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.763393] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1070.764093] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.764225] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.764558] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1070.765223] env[63355]: DEBUG oslo_concurrency.lockutils [None req-04791349-c186-490e-a703-2ee5eb9f89b2 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.252s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.766246] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa5f7b68-a905-4c26-a885-068188a734b9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.769355] env[63355]: DEBUG nova.network.neutron [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Successfully updated port: 2689b07f-15d6-4cad-b652-24c86298c985 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1070.771030] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.085s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.771466] env[63355]: DEBUG nova.compute.manager [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.773038] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26f3b5c-a0a9-4c5f-b97e-0c9abccd98a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.779515] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1070.779515] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526950fe-c295-9b89-0eb7-68a86321fc33" [ 1070.779515] env[63355]: _type = "Task" [ 1070.779515] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.782119] env[63355]: DEBUG nova.compute.manager [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63355) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1070.782772] env[63355]: DEBUG nova.objects.instance [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'flavor' on Instance uuid ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.796877] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526950fe-c295-9b89-0eb7-68a86321fc33, 'name': SearchDatastore_Task, 'duration_secs': 0.010388} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.797237] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.797607] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.797729] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.797867] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.798062] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.798548] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6489af0f-1f5f-446d-9d7d-e792c2ad22ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.806756] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.806756] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1070.807284] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4efa7a7e-d688-4f09-a5aa-6f2eb3acb324 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.812909] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1070.812909] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5234e31a-0be3-275b-bff4-ba85c2b5b87c" [ 1070.812909] env[63355]: _type = "Task" [ 1070.812909] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.820429] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5234e31a-0be3-275b-bff4-ba85c2b5b87c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.961614] env[63355]: DEBUG oslo_concurrency.lockutils [req-05bb7b11-3568-4f34-b012-364f182accd1 req-a450437b-fc49-4a7a-8d8a-b9034a9bc7f1 service nova] Releasing lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.983209] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350116, 'name': ReconfigVM_Task, 'duration_secs': 0.334569} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.983505] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Reconfigured VM instance instance-00000061 to attach disk [datastore2] f08bca27-ab50-40d0-a6d9-4842a37a4faf/f08bca27-ab50-40d0-a6d9-4842a37a4faf.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.984141] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ca279a26-1a8e-4499-aef8-64f268bac8ab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.990429] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1070.990429] env[63355]: value = "task-1350117" [ 1070.990429] env[63355]: _type = "Task" [ 1070.990429] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.999272] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350117, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.271997] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.272165] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.272313] env[63355]: DEBUG nova.network.neutron [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1071.292375] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1071.292375] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1642d9b0-5f5b-4ed7-8c8d-30992f269f59 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.298913] env[63355]: DEBUG oslo_vmware.api [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1071.298913] env[63355]: value = "task-1350118" [ 1071.298913] env[63355]: _type = "Task" [ 1071.298913] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.307316] env[63355]: DEBUG oslo_vmware.api [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350118, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.322266] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5234e31a-0be3-275b-bff4-ba85c2b5b87c, 'name': SearchDatastore_Task, 'duration_secs': 0.0081} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.322883] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-509e1a5a-ddac-4b2f-970e-c109e2b48ce9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.327849] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1071.327849] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522ea10c-b361-0808-350d-eb919d948603" [ 1071.327849] env[63355]: _type = "Task" [ 1071.327849] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.335655] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522ea10c-b361-0808-350d-eb919d948603, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.340708] env[63355]: DEBUG nova.network.neutron [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance_info_cache with network_info: [{"id": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "address": "fa:16:3e:8a:f8:58", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba0aaec1-3b", "ovs_interfaceid": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.500606] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350117, 'name': Rename_Task, 'duration_secs': 0.138907} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.500969] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1071.501139] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-036b62a3-52f1-49d3-8cf2-6e76701ff964 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.507088] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1071.507088] env[63355]: value = "task-1350119" [ 1071.507088] env[63355]: _type = "Task" [ 1071.507088] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.513917] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350119, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.588338] env[63355]: DEBUG nova.compute.manager [req-17f37b54-4490-4b0f-a7e9-24a59cea7bea req-69a40c71-541f-46c1-809d-695db04e1563 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Received event network-changed-2689b07f-15d6-4cad-b652-24c86298c985 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1071.588559] env[63355]: DEBUG nova.compute.manager [req-17f37b54-4490-4b0f-a7e9-24a59cea7bea req-69a40c71-541f-46c1-809d-695db04e1563 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Refreshing instance network info cache due to event network-changed-2689b07f-15d6-4cad-b652-24c86298c985. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1071.588750] env[63355]: DEBUG oslo_concurrency.lockutils [req-17f37b54-4490-4b0f-a7e9-24a59cea7bea req-69a40c71-541f-46c1-809d-695db04e1563 service nova] Acquiring lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.803695] env[63355]: DEBUG nova.network.neutron [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1071.812367] env[63355]: DEBUG oslo_vmware.api [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350118, 'name': PowerOffVM_Task, 'duration_secs': 0.177689} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.812714] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.812949] env[63355]: DEBUG nova.compute.manager [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1071.813814] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04052502-6241-42b8-a7a8-56b53e0bbefa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.838119] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522ea10c-b361-0808-350d-eb919d948603, 'name': SearchDatastore_Task, 'duration_secs': 0.009944} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.838390] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.838650] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 8fc3ea4e-d525-4583-b237-f895545fd3d1/8fc3ea4e-d525-4583-b237-f895545fd3d1.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1071.838916] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f84557b7-3e53-480f-832a-83be790afe4c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.843507] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.851289] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1071.851289] env[63355]: value = "task-1350120" [ 1071.851289] env[63355]: _type = "Task" [ 1071.851289] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.861483] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350120, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.877952] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.877952] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquired lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.877952] env[63355]: DEBUG nova.network.neutron [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1071.951576] env[63355]: DEBUG nova.network.neutron [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance_info_cache with network_info: [{"id": "2689b07f-15d6-4cad-b652-24c86298c985", "address": "fa:16:3e:cf:90:6f", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2689b07f-15", "ovs_interfaceid": "2689b07f-15d6-4cad-b652-24c86298c985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.016998] env[63355]: DEBUG oslo_vmware.api [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350119, 'name': PowerOnVM_Task, 'duration_secs': 0.435518} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.016998] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1072.017140] env[63355]: INFO nova.compute.manager [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Took 6.84 seconds to spawn the instance on the hypervisor. [ 1072.017410] env[63355]: DEBUG nova.compute.manager [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1072.018197] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8128a625-f226-4284-a7b8-a0cda3b54fd1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.289043] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "fa748748-31ec-431a-a628-5ea179e26fc8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.289315] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "fa748748-31ec-431a-a628-5ea179e26fc8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.289626] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "fa748748-31ec-431a-a628-5ea179e26fc8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.289863] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "fa748748-31ec-431a-a628-5ea179e26fc8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.290141] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "fa748748-31ec-431a-a628-5ea179e26fc8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.292800] env[63355]: INFO nova.compute.manager [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Terminating instance [ 1072.295097] env[63355]: DEBUG nova.compute.manager [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1072.295317] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1072.296152] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ebeee1-e19a-4523-8569-ba799c15c0df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.304564] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1072.304856] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59e26d18-4d9b-4f0e-b4f6-04bd81b4caf4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.312848] env[63355]: DEBUG oslo_vmware.api [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1072.312848] env[63355]: value = "task-1350121" [ 1072.312848] env[63355]: _type = "Task" [ 1072.312848] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.322846] env[63355]: DEBUG oslo_vmware.api [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.328338] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f657bb68-c509-4ae2-8f63-22b97be60d0f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.557s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.372168] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350120, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.381660] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c859b37b-afff-4b22-8edc-e7bcc52f4746 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.404638] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4209004c-211a-4143-a953-f9d445c44cf6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.411965] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance '50e7499c-a8c1-41a6-adac-36e6c3b92819' progress to 83 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1072.453882] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.454276] env[63355]: DEBUG nova.compute.manager [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Instance network_info: |[{"id": "2689b07f-15d6-4cad-b652-24c86298c985", "address": "fa:16:3e:cf:90:6f", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2689b07f-15", "ovs_interfaceid": "2689b07f-15d6-4cad-b652-24c86298c985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1072.454625] env[63355]: DEBUG oslo_concurrency.lockutils [req-17f37b54-4490-4b0f-a7e9-24a59cea7bea req-69a40c71-541f-46c1-809d-695db04e1563 service nova] Acquired lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.454842] env[63355]: DEBUG nova.network.neutron [req-17f37b54-4490-4b0f-a7e9-24a59cea7bea req-69a40c71-541f-46c1-809d-695db04e1563 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Refreshing network info cache for port 2689b07f-15d6-4cad-b652-24c86298c985 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1072.456303] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:90:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2689b07f-15d6-4cad-b652-24c86298c985', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1072.463809] env[63355]: DEBUG oslo.service.loopingcall [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.469587] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1072.470392] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e801bfe-9e81-457b-9492-c76f8dbded40 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.490320] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1072.490320] env[63355]: value = "task-1350122" [ 1072.490320] env[63355]: _type = "Task" [ 1072.490320] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.498395] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350122, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.540028] env[63355]: INFO nova.compute.manager [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Took 15.26 seconds to build instance. [ 1072.720566] env[63355]: INFO nova.network.neutron [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Port 563020c5-4b41-4820-9807-82cf864fcade from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1072.721311] env[63355]: DEBUG nova.network.neutron [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updating instance_info_cache with network_info: [{"id": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "address": "fa:16:3e:40:51:98", "network": {"id": "943d3ecf-0c77-4c51-a997-b17d86259c34", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2090178854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d47b671ea9c429391cbdae7e24adadf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap546aac3d-2d", "ovs_interfaceid": "546aac3d-2d71-48bd-84ac-6161e42a90dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.782910] env[63355]: DEBUG nova.network.neutron [req-17f37b54-4490-4b0f-a7e9-24a59cea7bea req-69a40c71-541f-46c1-809d-695db04e1563 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updated VIF entry in instance network info cache for port 2689b07f-15d6-4cad-b652-24c86298c985. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1072.783435] env[63355]: DEBUG nova.network.neutron [req-17f37b54-4490-4b0f-a7e9-24a59cea7bea req-69a40c71-541f-46c1-809d-695db04e1563 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance_info_cache with network_info: [{"id": "2689b07f-15d6-4cad-b652-24c86298c985", "address": "fa:16:3e:cf:90:6f", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2689b07f-15", "ovs_interfaceid": "2689b07f-15d6-4cad-b652-24c86298c985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.822482] env[63355]: DEBUG oslo_vmware.api [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350121, 'name': PowerOffVM_Task, 'duration_secs': 0.230183} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.822928] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1072.823133] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1072.823390] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fba5e160-3d53-465c-9555-76a02fb8a892 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.866233] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350120, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526083} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.866490] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 8fc3ea4e-d525-4583-b237-f895545fd3d1/8fc3ea4e-d525-4583-b237-f895545fd3d1.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1072.866705] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.866954] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d3720a8-3d55-4386-8e12-46a0051d57a5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.872749] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1072.872749] env[63355]: value = "task-1350124" [ 1072.872749] env[63355]: _type = "Task" [ 1072.872749] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.880881] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350124, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.889016] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1072.889249] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1072.889485] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Deleting the datastore file [datastore2] fa748748-31ec-431a-a628-5ea179e26fc8 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1072.889770] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edb5b2b2-7f46-4f5a-8257-1a18f11b6c05 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.895984] env[63355]: DEBUG oslo_vmware.api [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1072.895984] env[63355]: value = "task-1350125" [ 1072.895984] env[63355]: _type = "Task" [ 1072.895984] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.903776] env[63355]: DEBUG oslo_vmware.api [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350125, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.918868] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1072.919195] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8cd3ce4a-fd9f-48a5-854c-1f319f1051de {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.926343] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1072.926343] env[63355]: value = "task-1350126" [ 1072.926343] env[63355]: _type = "Task" [ 1072.926343] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.933766] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350126, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.999793] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350122, 'name': CreateVM_Task} progress is 25%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.020492] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.045238] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f29a0bc1-b4b8-47d5-bd30-607e1266e671 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.774s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.045564] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.025s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.045797] env[63355]: DEBUG nova.compute.manager [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1073.046682] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c654580-41b6-4d52-874c-dc20468b1794 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.053243] env[63355]: DEBUG nova.compute.manager [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63355) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1073.053806] env[63355]: DEBUG nova.objects.instance [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lazy-loading 'flavor' on Instance uuid f08bca27-ab50-40d0-a6d9-4842a37a4faf {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.073973] env[63355]: DEBUG nova.objects.instance [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'flavor' on Instance uuid ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.223913] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Releasing lock "refresh_cache-fa748748-31ec-431a-a628-5ea179e26fc8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.286702] env[63355]: DEBUG oslo_concurrency.lockutils [req-17f37b54-4490-4b0f-a7e9-24a59cea7bea req-69a40c71-541f-46c1-809d-695db04e1563 service nova] Releasing lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.382761] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350124, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093851} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.382991] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1073.383749] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c26709c-51bc-424e-a959-c0840bbd0946 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.406855] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 8fc3ea4e-d525-4583-b237-f895545fd3d1/8fc3ea4e-d525-4583-b237-f895545fd3d1.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.409882] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2593d440-22fe-4f3b-bd07-a9955367a806 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.428489] env[63355]: DEBUG oslo_vmware.api [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350125, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.380003} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.432197] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1073.432399] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1073.432579] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1073.432757] env[63355]: INFO nova.compute.manager [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1073.433007] env[63355]: DEBUG oslo.service.loopingcall [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.433268] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1073.433268] env[63355]: value = "task-1350127" [ 1073.433268] env[63355]: _type = "Task" [ 1073.433268] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.433465] env[63355]: DEBUG nova.compute.manager [-] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1073.433560] env[63355]: DEBUG nova.network.neutron [-] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1073.440991] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350126, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.445652] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350127, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.509448] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350122, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.560270] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1073.560558] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dad2546a-8369-4e1b-85e1-7ed2cc56d70f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.567089] env[63355]: DEBUG oslo_vmware.api [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1073.567089] env[63355]: value = "task-1350128" [ 1073.567089] env[63355]: _type = "Task" [ 1073.567089] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.575220] env[63355]: DEBUG oslo_vmware.api [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350128, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.578995] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.579222] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquired lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.579423] env[63355]: DEBUG nova.network.neutron [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1073.579609] env[63355]: DEBUG nova.objects.instance [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'info_cache' on Instance uuid ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.732035] env[63355]: DEBUG oslo_concurrency.lockutils [None req-69786f77-d5ac-46a9-85f9-94b4010c03a7 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "interface-fa748748-31ec-431a-a628-5ea179e26fc8-563020c5-4b41-4820-9807-82cf864fcade" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.876s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.938815] env[63355]: DEBUG oslo_vmware.api [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350126, 'name': PowerOnVM_Task, 'duration_secs': 0.863968} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.941893] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1073.942108] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8c638ade-425b-4843-aaa1-fd44431e5606 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance '50e7499c-a8c1-41a6-adac-36e6c3b92819' progress to 100 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1073.951257] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350127, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.003394] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350122, 'name': CreateVM_Task, 'duration_secs': 1.135967} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.003913] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1074.004280] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.004470] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.004754] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1074.005019] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a31b4684-90c0-4bd2-80b8-f4a183102370 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.010299] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1074.010299] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ce70f4-ccc9-c888-eb7a-237b0d437923" [ 1074.010299] env[63355]: _type = "Task" [ 1074.010299] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.018406] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ce70f4-ccc9-c888-eb7a-237b0d437923, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.076331] env[63355]: DEBUG oslo_vmware.api [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350128, 'name': PowerOffVM_Task, 'duration_secs': 0.235858} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.076632] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1074.076832] env[63355]: DEBUG nova.compute.manager [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.077690] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93349ed1-18b8-4f74-9aa3-e998e175073a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.082862] env[63355]: DEBUG nova.objects.base [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1074.128012] env[63355]: DEBUG nova.compute.manager [req-0dcb1d68-1cd1-43b6-bf27-c2ca322ced87 req-87fceaee-5a51-4c7c-9e40-277c13146e55 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Received event network-vif-deleted-546aac3d-2d71-48bd-84ac-6161e42a90dd {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1074.128309] env[63355]: INFO nova.compute.manager [req-0dcb1d68-1cd1-43b6-bf27-c2ca322ced87 req-87fceaee-5a51-4c7c-9e40-277c13146e55 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Neutron deleted interface 546aac3d-2d71-48bd-84ac-6161e42a90dd; detaching it from the instance and deleting it from the info cache [ 1074.128472] env[63355]: DEBUG nova.network.neutron [req-0dcb1d68-1cd1-43b6-bf27-c2ca322ced87 req-87fceaee-5a51-4c7c-9e40-277c13146e55 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.451248] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350127, 'name': ReconfigVM_Task, 'duration_secs': 0.565326} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.451529] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 8fc3ea4e-d525-4583-b237-f895545fd3d1/8fc3ea4e-d525-4583-b237-f895545fd3d1.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.452110] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8db7fe8e-e1c3-4fe5-9aa4-76a15cfb715e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.459379] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1074.459379] env[63355]: value = "task-1350129" [ 1074.459379] env[63355]: _type = "Task" [ 1074.459379] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.469490] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350129, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.519046] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ce70f4-ccc9-c888-eb7a-237b0d437923, 'name': SearchDatastore_Task, 'duration_secs': 0.046896} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.519355] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.519616] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1074.519864] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.520028] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.520218] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1074.520462] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0734ec20-809e-443a-bb81-3e7127949689 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.528178] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1074.528374] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1074.529029] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93fb3918-8971-462f-8ca2-f98d97f1a653 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.533737] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1074.533737] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c915b1-14c2-9f18-326d-61a79f8346b7" [ 1074.533737] env[63355]: _type = "Task" [ 1074.533737] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.540682] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c915b1-14c2-9f18-326d-61a79f8346b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.593325] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5310b3cc-be36-43b0-9776-ca06e54fa753 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.607088] env[63355]: DEBUG nova.network.neutron [-] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.631182] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa0b5906-2b0e-4aa7-b33e-3286247ee9e0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.640760] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a9dcd3-09fd-4109-b266-302a045a4077 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.667222] env[63355]: DEBUG nova.compute.manager [req-0dcb1d68-1cd1-43b6-bf27-c2ca322ced87 req-87fceaee-5a51-4c7c-9e40-277c13146e55 service nova] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Detach interface failed, port_id=546aac3d-2d71-48bd-84ac-6161e42a90dd, reason: Instance fa748748-31ec-431a-a628-5ea179e26fc8 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1074.796745] env[63355]: DEBUG nova.network.neutron [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updating instance_info_cache with network_info: [{"id": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "address": "fa:16:3e:fb:af:dc", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4c4170a-d2", "ovs_interfaceid": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.970464] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350129, 'name': Rename_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.045230] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52c915b1-14c2-9f18-326d-61a79f8346b7, 'name': SearchDatastore_Task, 'duration_secs': 0.067896} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.045987] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b53d2f5-0bb5-4f8f-8025-ad0ad34ef6b7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.051741] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1075.051741] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a80ac5-4f2f-bbad-b2b3-0229230d0500" [ 1075.051741] env[63355]: _type = "Task" [ 1075.051741] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.059751] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a80ac5-4f2f-bbad-b2b3-0229230d0500, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.110625] env[63355]: INFO nova.compute.manager [-] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Took 1.68 seconds to deallocate network for instance. [ 1075.302979] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Releasing lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.476890] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350129, 'name': Rename_Task, 'duration_secs': 0.704599} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.478485] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1075.478881] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-161aa2c3-e3a7-4896-9b76-e6d178382151 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.484811] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1075.484811] env[63355]: value = "task-1350130" [ 1075.484811] env[63355]: _type = "Task" [ 1075.484811] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.493599] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350130, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.563358] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a80ac5-4f2f-bbad-b2b3-0229230d0500, 'name': SearchDatastore_Task, 'duration_secs': 0.013954} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.563634] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.563893] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50/8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1075.564162] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4c0d16e-711f-4b3b-8bbf-291472c9317d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.570219] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1075.570219] env[63355]: value = "task-1350131" [ 1075.570219] env[63355]: _type = "Task" [ 1075.570219] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.577629] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350131, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.617050] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.617436] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.617641] env[63355]: DEBUG nova.objects.instance [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'resources' on Instance uuid fa748748-31ec-431a-a628-5ea179e26fc8 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.727525] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.727780] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.728024] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.728220] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.728440] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.731115] env[63355]: INFO nova.compute.manager [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Terminating instance [ 1075.732916] env[63355]: DEBUG nova.compute.manager [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1075.733125] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1075.733947] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee6b565-e052-4b81-9cc2-fde0540315c8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.741497] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1075.741741] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af7157f8-c1a7-4740-b33a-882fc85aedc1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.805955] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1075.806423] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df3193a3-363c-44ff-87ff-0ae8078b1874 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.815637] env[63355]: DEBUG oslo_vmware.api [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1075.815637] env[63355]: value = "task-1350133" [ 1075.815637] env[63355]: _type = "Task" [ 1075.815637] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.826103] env[63355]: DEBUG oslo_vmware.api [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350133, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.827571] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1075.827796] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1075.827979] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleting the datastore file [datastore2] f08bca27-ab50-40d0-a6d9-4842a37a4faf {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1075.828276] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94d70969-6ba6-45c5-958f-2580a95a17a3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.834246] env[63355]: DEBUG oslo_vmware.api [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for the task: (returnval){ [ 1075.834246] env[63355]: value = "task-1350134" [ 1075.834246] env[63355]: _type = "Task" [ 1075.834246] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.842382] env[63355]: DEBUG oslo_vmware.api [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350134, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.995492] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350130, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.087557] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350131, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.252349] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16216794-3996-49e9-9e52-c16ec7c8dbb3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.259717] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fe558a-b7b2-4c8b-8f09-a588b7ebc358 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.289732] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa9185a-81f9-48f4-b836-ca1b304b0c46 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.297363] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fdf749-046f-46de-9635-d7f52bfd7864 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.310404] env[63355]: DEBUG nova.compute.provider_tree [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.324195] env[63355]: DEBUG oslo_vmware.api [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350133, 'name': PowerOnVM_Task, 'duration_secs': 0.462791} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.324443] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1076.324631] env[63355]: DEBUG nova.compute.manager [None req-2f2dae0d-38ef-4202-b4b3-ebf72d6c54b6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1076.325412] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32703e6-1020-4892-a28d-b891763b3d9f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.341955] env[63355]: DEBUG oslo_vmware.api [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Task: {'id': task-1350134, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.279272} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.342204] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1076.342392] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1076.342567] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1076.342763] env[63355]: INFO nova.compute.manager [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1076.343970] env[63355]: DEBUG oslo.service.loopingcall [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.343970] env[63355]: DEBUG nova.compute.manager [-] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1076.343970] env[63355]: DEBUG nova.network.neutron [-] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1076.494793] env[63355]: DEBUG oslo_vmware.api [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350130, 'name': PowerOnVM_Task, 'duration_secs': 0.552659} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.495088] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1076.495297] env[63355]: INFO nova.compute.manager [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Took 8.20 seconds to spawn the instance on the hypervisor. [ 1076.495480] env[63355]: DEBUG nova.compute.manager [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1076.496241] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e7f3bb-b753-4346-a863-3f31f214fce6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.581819] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350131, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539448} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.582334] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50/8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1076.582562] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1076.582815] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1a4a8e6a-d52b-4692-a803-ca2a89c5ea7d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.592178] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1076.592178] env[63355]: value = "task-1350135" [ 1076.592178] env[63355]: _type = "Task" [ 1076.592178] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.600121] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350135, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.602074] env[63355]: DEBUG nova.compute.manager [req-d8233762-4638-45c0-adb0-44d73123802f req-4982ef29-d580-47a8-a7d6-377bbac60880 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Received event network-vif-deleted-9b26dcfe-3e4b-4d6c-bd1f-c628bd055377 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1076.602320] env[63355]: INFO nova.compute.manager [req-d8233762-4638-45c0-adb0-44d73123802f req-4982ef29-d580-47a8-a7d6-377bbac60880 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Neutron deleted interface 9b26dcfe-3e4b-4d6c-bd1f-c628bd055377; detaching it from the instance and deleting it from the info cache [ 1076.602504] env[63355]: DEBUG nova.network.neutron [req-d8233762-4638-45c0-adb0-44d73123802f req-4982ef29-d580-47a8-a7d6-377bbac60880 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.685605] env[63355]: DEBUG oslo_concurrency.lockutils [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "50e7499c-a8c1-41a6-adac-36e6c3b92819" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.685916] env[63355]: DEBUG oslo_concurrency.lockutils [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.686121] env[63355]: DEBUG nova.compute.manager [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Going to confirm migration 3 {{(pid=63355) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1076.813217] env[63355]: DEBUG nova.scheduler.client.report [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1077.012756] env[63355]: INFO nova.compute.manager [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Took 12.98 seconds to build instance. [ 1077.083173] env[63355]: DEBUG nova.network.neutron [-] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.104068] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350135, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.104711] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e575e942-85d8-42bf-b032-543052fb9488 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.114591] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca32227-453d-4f46-801a-e547355fd29f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.142214] env[63355]: DEBUG nova.compute.manager [req-d8233762-4638-45c0-adb0-44d73123802f req-4982ef29-d580-47a8-a7d6-377bbac60880 service nova] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Detach interface failed, port_id=9b26dcfe-3e4b-4d6c-bd1f-c628bd055377, reason: Instance f08bca27-ab50-40d0-a6d9-4842a37a4faf could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1077.266107] env[63355]: DEBUG oslo_concurrency.lockutils [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.266320] env[63355]: DEBUG oslo_concurrency.lockutils [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.266503] env[63355]: DEBUG nova.network.neutron [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1077.266693] env[63355]: DEBUG nova.objects.instance [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lazy-loading 'info_cache' on Instance uuid 50e7499c-a8c1-41a6-adac-36e6c3b92819 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.317825] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.339259] env[63355]: INFO nova.scheduler.client.report [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Deleted allocations for instance fa748748-31ec-431a-a628-5ea179e26fc8 [ 1077.514958] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b656b1e0-dad9-47d1-a7aa-f3b8b9d4b8f1 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.488s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.585463] env[63355]: INFO nova.compute.manager [-] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Took 1.24 seconds to deallocate network for instance. [ 1077.602133] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350135, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.857353} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.602482] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1077.603247] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7690cdd5-ab9e-4f15-b927-61c7c6e3ecf4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.624733] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50/8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1077.625556] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54fd7d9b-bfc7-4e2b-b130-71c949a84272 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.644929] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1077.644929] env[63355]: value = "task-1350136" [ 1077.644929] env[63355]: _type = "Task" [ 1077.644929] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.655831] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350136, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.847668] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8e80a7ca-0c95-44f2-a0e9-fd2d09ad8007 tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "fa748748-31ec-431a-a628-5ea179e26fc8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.558s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.889923] env[63355]: DEBUG nova.compute.manager [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Stashing vm_state: active {{(pid=63355) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1078.093056] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.093395] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.093658] env[63355]: DEBUG nova.objects.instance [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lazy-loading 'resources' on Instance uuid f08bca27-ab50-40d0-a6d9-4842a37a4faf {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.154670] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.430220] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.529491] env[63355]: DEBUG nova.network.neutron [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance_info_cache with network_info: [{"id": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "address": "fa:16:3e:8a:f8:58", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba0aaec1-3b", "ovs_interfaceid": "ba0aaec1-3b3e-4594-bf65-aa18843e439d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.655577] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350136, 'name': ReconfigVM_Task, 'duration_secs': 0.862247} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.658301] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50/8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.659745] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57f886ce-dad6-41fd-a9a0-137c077da610 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.665433] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1078.665433] env[63355]: value = "task-1350137" [ 1078.665433] env[63355]: _type = "Task" [ 1078.665433] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.676103] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350137, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.724389] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3c6e24-e46a-492b-92d4-d4c1c2872c5f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.732159] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65acaf9c-b703-429e-9a79-a7543ca09bf5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.774366] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc62a577-14c8-41fc-b613-7bf9b534eec9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.784845] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb61d0c1-b983-46f8-b96b-b2743411ef54 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.798218] env[63355]: DEBUG nova.compute.provider_tree [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.838459] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.838817] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.839074] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "da3cb83c-6368-49c0-9b11-0498221e3c0f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.839326] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "da3cb83c-6368-49c0-9b11-0498221e3c0f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.839618] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "da3cb83c-6368-49c0-9b11-0498221e3c0f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.841808] env[63355]: INFO nova.compute.manager [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Terminating instance [ 1078.843670] env[63355]: DEBUG nova.compute.manager [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1078.843884] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1078.844754] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c6b809-47ac-4c11-8b33-883d08364f8f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.854159] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1078.854477] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b47a8529-3ae7-4b6b-9c13-1d49938c6754 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.860601] env[63355]: DEBUG oslo_vmware.api [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1078.860601] env[63355]: value = "task-1350138" [ 1078.860601] env[63355]: _type = "Task" [ 1078.860601] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.867973] env[63355]: DEBUG oslo_vmware.api [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.034032] env[63355]: DEBUG oslo_concurrency.lockutils [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-50e7499c-a8c1-41a6-adac-36e6c3b92819" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.034032] env[63355]: DEBUG nova.objects.instance [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lazy-loading 'migration_context' on Instance uuid 50e7499c-a8c1-41a6-adac-36e6c3b92819 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.175674] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350137, 'name': Rename_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.301100] env[63355]: DEBUG nova.scheduler.client.report [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.370559] env[63355]: DEBUG oslo_vmware.api [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.536196] env[63355]: DEBUG nova.objects.base [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Object Instance<50e7499c-a8c1-41a6-adac-36e6c3b92819> lazy-loaded attributes: info_cache,migration_context {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1079.537146] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0ff7ac-5998-49fc-8297-2759eb116f2e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.556062] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af6628dd-148f-4cfc-bbd1-df74079cefe2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.561376] env[63355]: DEBUG oslo_vmware.api [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1079.561376] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528ac104-524d-b89b-795d-99607e3bdf55" [ 1079.561376] env[63355]: _type = "Task" [ 1079.561376] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.568856] env[63355]: DEBUG oslo_vmware.api [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528ac104-524d-b89b-795d-99607e3bdf55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.675889] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350137, 'name': Rename_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.806290] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.713s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.808640] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.379s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.826364] env[63355]: INFO nova.scheduler.client.report [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Deleted allocations for instance f08bca27-ab50-40d0-a6d9-4842a37a4faf [ 1079.871819] env[63355]: DEBUG oslo_vmware.api [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350138, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.072987] env[63355]: DEBUG oslo_vmware.api [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]528ac104-524d-b89b-795d-99607e3bdf55, 'name': SearchDatastore_Task, 'duration_secs': 0.013489} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.073311] env[63355]: DEBUG oslo_concurrency.lockutils [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.175794] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350137, 'name': Rename_Task, 'duration_secs': 1.151626} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.177055] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.177055] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d48dddb2-a0b0-4284-9ffe-a6836c123a6e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.182152] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1080.182152] env[63355]: value = "task-1350139" [ 1080.182152] env[63355]: _type = "Task" [ 1080.182152] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.190315] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350139, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.313489] env[63355]: INFO nova.compute.claims [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1080.334721] env[63355]: DEBUG oslo_concurrency.lockutils [None req-1530f19b-ac6e-4c4c-a09a-0d4c70ee0b89 tempest-ServersTestJSON-549581366 tempest-ServersTestJSON-549581366-project-member] Lock "f08bca27-ab50-40d0-a6d9-4842a37a4faf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.607s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.371607] env[63355]: DEBUG oslo_vmware.api [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350138, 'name': PowerOffVM_Task, 'duration_secs': 1.085875} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.371866] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.372050] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1080.372301] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fac2ff70-9e64-4919-9578-579f31074f87 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.470862] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1080.471103] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1080.471297] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Deleting the datastore file [datastore1] da3cb83c-6368-49c0-9b11-0498221e3c0f {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.471581] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0548e9f7-4877-4001-9b02-88bd64128385 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.478653] env[63355]: DEBUG oslo_vmware.api [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for the task: (returnval){ [ 1080.478653] env[63355]: value = "task-1350141" [ 1080.478653] env[63355]: _type = "Task" [ 1080.478653] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.486737] env[63355]: DEBUG oslo_vmware.api [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350141, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.692440] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350139, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.821113] env[63355]: INFO nova.compute.resource_tracker [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating resource usage from migration a01d617f-dd26-4832-8207-3e966623a221 [ 1080.929036] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a5af43-b518-4672-a5b2-c5be40d608d1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.940425] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf2c364-850c-4f6e-ba42-af017e93c774 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.977366] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61c012a-4706-457d-badd-6297a006d143 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.990783] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7f29e5-8387-40b4-a2ef-4ea07fc850ab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.994445] env[63355]: DEBUG oslo_vmware.api [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Task: {'id': task-1350141, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172212} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.994987] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1080.995255] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1080.995457] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1080.995636] env[63355]: INFO nova.compute.manager [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Took 2.15 seconds to destroy the instance on the hypervisor. [ 1080.995871] env[63355]: DEBUG oslo.service.loopingcall [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1080.996442] env[63355]: DEBUG nova.compute.manager [-] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1080.996539] env[63355]: DEBUG nova.network.neutron [-] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1081.008629] env[63355]: DEBUG nova.compute.provider_tree [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.194729] env[63355]: DEBUG oslo_vmware.api [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350139, 'name': PowerOnVM_Task, 'duration_secs': 0.567455} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.194729] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.194729] env[63355]: INFO nova.compute.manager [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Took 10.70 seconds to spawn the instance on the hypervisor. [ 1081.194729] env[63355]: DEBUG nova.compute.manager [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1081.195075] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b415ef-2e3d-4021-842a-a661abad53f3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.307699] env[63355]: DEBUG nova.compute.manager [req-75fb1dd2-2e5f-48eb-908a-041f9c9b0102 req-001db3c9-0139-42c9-938a-6df0bb0d292e service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Received event network-vif-deleted-aa5a40b6-5d12-4d62-88d2-32a26a65a241 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.307918] env[63355]: INFO nova.compute.manager [req-75fb1dd2-2e5f-48eb-908a-041f9c9b0102 req-001db3c9-0139-42c9-938a-6df0bb0d292e service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Neutron deleted interface aa5a40b6-5d12-4d62-88d2-32a26a65a241; detaching it from the instance and deleting it from the info cache [ 1081.308111] env[63355]: DEBUG nova.network.neutron [req-75fb1dd2-2e5f-48eb-908a-041f9c9b0102 req-001db3c9-0139-42c9-938a-6df0bb0d292e service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.436956] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.437301] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.437535] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.437832] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.438018] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.440201] env[63355]: INFO nova.compute.manager [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Terminating instance [ 1081.442093] env[63355]: DEBUG nova.compute.manager [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1081.442255] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1081.443164] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6febd929-9b00-4e94-bdfe-df0cea442b50 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.452416] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1081.452625] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f980ceb8-316c-4554-b7b4-03b4e6142d68 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.458790] env[63355]: DEBUG oslo_vmware.api [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1081.458790] env[63355]: value = "task-1350142" [ 1081.458790] env[63355]: _type = "Task" [ 1081.458790] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.466348] env[63355]: DEBUG oslo_vmware.api [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.511702] env[63355]: DEBUG nova.scheduler.client.report [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.712184] env[63355]: INFO nova.compute.manager [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Took 15.66 seconds to build instance. [ 1081.783219] env[63355]: DEBUG nova.network.neutron [-] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.813282] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3ad04a3-5848-472e-9bae-7d1c597255c7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.820436] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b22fae-4cbb-4b0a-8286-f2a93d2e5bd2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.850967] env[63355]: DEBUG nova.compute.manager [req-75fb1dd2-2e5f-48eb-908a-041f9c9b0102 req-001db3c9-0139-42c9-938a-6df0bb0d292e service nova] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Detach interface failed, port_id=aa5a40b6-5d12-4d62-88d2-32a26a65a241, reason: Instance da3cb83c-6368-49c0-9b11-0498221e3c0f could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1081.969133] env[63355]: DEBUG oslo_vmware.api [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350142, 'name': PowerOffVM_Task, 'duration_secs': 0.2005} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.969958] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1081.970329] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1081.970707] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dfdec67b-05b4-4618-8d95-26170527acaa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.016582] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.208s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.017154] env[63355]: INFO nova.compute.manager [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Migrating [ 1082.025537] env[63355]: DEBUG oslo_concurrency.lockutils [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.951s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.214345] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c8fc3135-6dc1-4fb1-9769-fc0fb084cf1c tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.174s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.259141] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1082.259538] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1082.259971] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Deleting the datastore file [datastore2] 7c153109-b814-4e11-b4f9-7b8cebb853d4 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1082.261026] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de9d64f8-66db-46a0-b9b2-4bbbdb7ab984 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.268717] env[63355]: DEBUG oslo_vmware.api [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for the task: (returnval){ [ 1082.268717] env[63355]: value = "task-1350144" [ 1082.268717] env[63355]: _type = "Task" [ 1082.268717] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.280614] env[63355]: DEBUG oslo_vmware.api [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350144, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.287031] env[63355]: INFO nova.compute.manager [-] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Took 1.29 seconds to deallocate network for instance. [ 1082.535749] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.535957] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.536162] env[63355]: DEBUG nova.network.neutron [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1082.652707] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d02984e-9784-48a9-aaf6-dfda5ca34fca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.659599] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e339dbe-17d9-413b-9a13-bfe96f5ea8cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.689235] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6bc1a77-64c7-40cd-8b12-617452c77edb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.696695] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d470d0-5762-490c-9d3e-3ae4ad5714df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.711428] env[63355]: DEBUG nova.compute.provider_tree [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.777726] env[63355]: DEBUG oslo_vmware.api [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Task: {'id': task-1350144, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210778} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.777999] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1082.778209] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1082.778390] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1082.778569] env[63355]: INFO nova.compute.manager [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Took 1.34 seconds to destroy the instance on the hypervisor. [ 1082.778851] env[63355]: DEBUG oslo.service.loopingcall [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1082.779029] env[63355]: DEBUG nova.compute.manager [-] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1082.779112] env[63355]: DEBUG nova.network.neutron [-] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1082.794801] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.215244] env[63355]: DEBUG nova.scheduler.client.report [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.345297] env[63355]: DEBUG nova.compute.manager [req-556a4eb0-4b5b-413b-8199-16e314c90328 req-64d45310-bb5f-4575-becb-9bd4c682d534 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Received event network-changed-2689b07f-15d6-4cad-b652-24c86298c985 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.345501] env[63355]: DEBUG nova.compute.manager [req-556a4eb0-4b5b-413b-8199-16e314c90328 req-64d45310-bb5f-4575-becb-9bd4c682d534 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Refreshing instance network info cache due to event network-changed-2689b07f-15d6-4cad-b652-24c86298c985. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1083.345724] env[63355]: DEBUG oslo_concurrency.lockutils [req-556a4eb0-4b5b-413b-8199-16e314c90328 req-64d45310-bb5f-4575-becb-9bd4c682d534 service nova] Acquiring lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.345874] env[63355]: DEBUG oslo_concurrency.lockutils [req-556a4eb0-4b5b-413b-8199-16e314c90328 req-64d45310-bb5f-4575-becb-9bd4c682d534 service nova] Acquired lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.346069] env[63355]: DEBUG nova.network.neutron [req-556a4eb0-4b5b-413b-8199-16e314c90328 req-64d45310-bb5f-4575-becb-9bd4c682d534 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Refreshing network info cache for port 2689b07f-15d6-4cad-b652-24c86298c985 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1083.576521] env[63355]: DEBUG nova.network.neutron [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance_info_cache with network_info: [{"id": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "address": "fa:16:3e:01:27:93", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca66441-d9", "ovs_interfaceid": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.060071] env[63355]: DEBUG nova.network.neutron [-] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.078806] env[63355]: DEBUG nova.network.neutron [req-556a4eb0-4b5b-413b-8199-16e314c90328 req-64d45310-bb5f-4575-becb-9bd4c682d534 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updated VIF entry in instance network info cache for port 2689b07f-15d6-4cad-b652-24c86298c985. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1084.079432] env[63355]: DEBUG nova.network.neutron [req-556a4eb0-4b5b-413b-8199-16e314c90328 req-64d45310-bb5f-4575-becb-9bd4c682d534 service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance_info_cache with network_info: [{"id": "2689b07f-15d6-4cad-b652-24c86298c985", "address": "fa:16:3e:cf:90:6f", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2689b07f-15", "ovs_interfaceid": "2689b07f-15d6-4cad-b652-24c86298c985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.080965] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.227551] env[63355]: DEBUG oslo_concurrency.lockutils [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.203s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.230613] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.438s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.230851] env[63355]: DEBUG nova.objects.instance [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lazy-loading 'resources' on Instance uuid da3cb83c-6368-49c0-9b11-0498221e3c0f {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.562453] env[63355]: INFO nova.compute.manager [-] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Took 1.78 seconds to deallocate network for instance. [ 1084.583381] env[63355]: DEBUG oslo_concurrency.lockutils [req-556a4eb0-4b5b-413b-8199-16e314c90328 req-64d45310-bb5f-4575-becb-9bd4c682d534 service nova] Releasing lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.795168] env[63355]: INFO nova.scheduler.client.report [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted allocation for migration bc6e8ef9-18e4-43c9-a2db-1b629c809422 [ 1084.856524] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4547ee-8efe-4453-afc1-56c7bf1e827a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.864827] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fa2c79-5167-4bd3-be49-34331551ec66 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.895585] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f92eaf-2798-43d2-ad2d-cb366fe567e8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.903498] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b483473-50ae-41b7-8107-ff95f87c4521 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.916948] env[63355]: DEBUG nova.compute.provider_tree [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.068744] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.235546] env[63355]: DEBUG oslo_concurrency.lockutils [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "50e7499c-a8c1-41a6-adac-36e6c3b92819" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.302286] env[63355]: DEBUG oslo_concurrency.lockutils [None req-435e780b-c2cc-4388-b333-4946071b8411 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.616s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.303487] env[63355]: DEBUG oslo_concurrency.lockutils [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.068s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.303728] env[63355]: DEBUG oslo_concurrency.lockutils [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "50e7499c-a8c1-41a6-adac-36e6c3b92819-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.303936] env[63355]: DEBUG oslo_concurrency.lockutils [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.304186] env[63355]: DEBUG oslo_concurrency.lockutils [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.306074] env[63355]: INFO nova.compute.manager [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Terminating instance [ 1085.307983] env[63355]: DEBUG nova.compute.manager [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1085.308205] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1085.309081] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60cc0d79-74d5-4870-aba1-dbfd424c0f17 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.317163] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1085.317418] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e5730c7b-2bbd-4b64-a2b9-4802c2cac573 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.324862] env[63355]: DEBUG oslo_vmware.api [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1085.324862] env[63355]: value = "task-1350145" [ 1085.324862] env[63355]: _type = "Task" [ 1085.324862] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.334218] env[63355]: DEBUG oslo_vmware.api [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350145, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.366956] env[63355]: DEBUG nova.compute.manager [req-0dbf14c5-b28f-4f31-b802-1f8ae62174af req-019e5b6b-e273-4824-b515-8f0c76cfd85c service nova] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Received event network-vif-deleted-43ebe841-d4c8-446a-981c-519bef977228 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1085.420326] env[63355]: DEBUG nova.scheduler.client.report [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1085.597830] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1b7230-44d7-459d-8620-cbd9f657b6cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.617769] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance '8fc3ea4e-d525-4583-b237-f895545fd3d1' progress to 0 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1085.695765] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "bcf935ad-afa2-4b36-b01b-14020453ad98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.696016] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "bcf935ad-afa2-4b36-b01b-14020453ad98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.834520] env[63355]: DEBUG oslo_vmware.api [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350145, 'name': PowerOffVM_Task, 'duration_secs': 0.228927} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.834806] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1085.834977] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1085.835260] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4069bef-46b1-4c11-97ba-037b3831ae6f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.899273] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1085.899563] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1085.899936] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleting the datastore file [datastore1] 50e7499c-a8c1-41a6-adac-36e6c3b92819 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1085.900229] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e434af79-6afa-4faa-809d-2a0e7107b333 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.906507] env[63355]: DEBUG oslo_vmware.api [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1085.906507] env[63355]: value = "task-1350147" [ 1085.906507] env[63355]: _type = "Task" [ 1085.906507] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.916393] env[63355]: DEBUG oslo_vmware.api [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350147, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.925323] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.695s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.927395] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.859s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.927634] env[63355]: DEBUG nova.objects.instance [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lazy-loading 'resources' on Instance uuid 7c153109-b814-4e11-b4f9-7b8cebb853d4 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.942685] env[63355]: INFO nova.scheduler.client.report [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Deleted allocations for instance da3cb83c-6368-49c0-9b11-0498221e3c0f [ 1086.124061] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1086.124756] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a1bb1e5e-3e47-4660-beaf-06e7149e4068 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.131878] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1086.131878] env[63355]: value = "task-1350148" [ 1086.131878] env[63355]: _type = "Task" [ 1086.131878] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.139465] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.198215] env[63355]: DEBUG nova.compute.manager [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1086.417445] env[63355]: DEBUG oslo_vmware.api [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350147, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220309} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.417789] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1086.417855] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1086.418043] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1086.418224] env[63355]: INFO nova.compute.manager [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1086.418511] env[63355]: DEBUG oslo.service.loopingcall [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1086.418704] env[63355]: DEBUG nova.compute.manager [-] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1086.418801] env[63355]: DEBUG nova.network.neutron [-] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1086.459873] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8586a586-6353-4e4f-83c6-a41aa264c00d tempest-AttachInterfacesTestJSON-1420495164 tempest-AttachInterfacesTestJSON-1420495164-project-member] Lock "da3cb83c-6368-49c0-9b11-0498221e3c0f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.621s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.525078] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebbc03a4-f5ba-4080-bd4b-11356a9ae331 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.532448] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35615c21-2684-4273-892c-f8f397ccb9b9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.563368] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e94fe9-2037-4c53-bab6-26917982a4ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.571126] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0844a9f6-b324-4716-9e6b-1417fc048479 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.584947] env[63355]: DEBUG nova.compute.provider_tree [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.641593] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350148, 'name': PowerOffVM_Task, 'duration_secs': 0.197069} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.641877] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1086.642086] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance '8fc3ea4e-d525-4583-b237-f895545fd3d1' progress to 17 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1086.720435] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.087895] env[63355]: DEBUG nova.scheduler.client.report [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.149014] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1087.149298] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1087.149467] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1087.149677] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1087.149831] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1087.149986] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1087.150329] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1087.150505] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1087.150681] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1087.150851] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1087.151042] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1087.155978] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c1d7c1a-6be8-45c2-999d-ab6336529059 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.172043] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1087.172043] env[63355]: value = "task-1350149" [ 1087.172043] env[63355]: _type = "Task" [ 1087.172043] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.183266] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350149, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.222058] env[63355]: DEBUG nova.network.neutron [-] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.390723] env[63355]: DEBUG nova.compute.manager [req-1214f08d-b362-495a-89a3-8828f72b0d8d req-9dea8cc0-c9b1-402b-a49c-34bdd2d17833 service nova] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Received event network-vif-deleted-ba0aaec1-3b3e-4594-bf65-aa18843e439d {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.593643] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.595919] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.876s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.597425] env[63355]: INFO nova.compute.claims [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.614802] env[63355]: INFO nova.scheduler.client.report [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Deleted allocations for instance 7c153109-b814-4e11-b4f9-7b8cebb853d4 [ 1087.685539] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350149, 'name': ReconfigVM_Task, 'duration_secs': 0.197796} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.685873] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance '8fc3ea4e-d525-4583-b237-f895545fd3d1' progress to 33 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1087.724861] env[63355]: INFO nova.compute.manager [-] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Took 1.31 seconds to deallocate network for instance. [ 1088.122545] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ec057adf-66d1-424f-9757-452a3e67728f tempest-AttachVolumeShelveTestJSON-199973924 tempest-AttachVolumeShelveTestJSON-199973924-project-member] Lock "7c153109-b814-4e11-b4f9-7b8cebb853d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.685s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.195736] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1088.196028] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1088.196200] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1088.196388] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1088.196713] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1088.196713] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1088.196883] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1088.197070] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1088.197230] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1088.197459] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1088.197609] env[63355]: DEBUG nova.virt.hardware [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1088.202966] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Reconfiguring VM instance instance-00000062 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1088.203597] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-409d3991-6355-4f5d-9e57-ad6989cbeab9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.223650] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1088.223650] env[63355]: value = "task-1350150" [ 1088.223650] env[63355]: _type = "Task" [ 1088.223650] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.232868] env[63355]: DEBUG oslo_concurrency.lockutils [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.233143] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350150, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.690461] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb1ccf9c-1a5f-4c38-9fb5-e942ae730e3e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.698290] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aad8d14-c170-4037-b942-a628b5853414 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.731358] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace009b2-ef70-4818-a740-82baae4533d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.738904] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350150, 'name': ReconfigVM_Task, 'duration_secs': 0.16149} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.740859] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Reconfigured VM instance instance-00000062 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1088.741633] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f331e0e1-0ac6-434e-8b9e-7ae3528e50b4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.745310] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-747a6c36-c26d-4983-b546-e2d805827c7f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.758537] env[63355]: DEBUG nova.compute.provider_tree [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.776128] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 8fc3ea4e-d525-4583-b237-f895545fd3d1/8fc3ea4e-d525-4583-b237-f895545fd3d1.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.777117] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca612425-0570-4930-98bc-34747d093c24 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.795424] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1088.795424] env[63355]: value = "task-1350151" [ 1088.795424] env[63355]: _type = "Task" [ 1088.795424] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.802879] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350151, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.277850] env[63355]: DEBUG nova.scheduler.client.report [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1089.305273] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350151, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.783346] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.187s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.783886] env[63355]: DEBUG nova.compute.manager [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1089.786538] env[63355]: DEBUG oslo_concurrency.lockutils [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.554s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.787816] env[63355]: DEBUG oslo_concurrency.lockutils [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.805942] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350151, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.806858] env[63355]: INFO nova.scheduler.client.report [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted allocations for instance 50e7499c-a8c1-41a6-adac-36e6c3b92819 [ 1090.290851] env[63355]: DEBUG nova.compute.utils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1090.292311] env[63355]: DEBUG nova.compute.manager [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1090.292482] env[63355]: DEBUG nova.network.neutron [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1090.307639] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350151, 'name': ReconfigVM_Task, 'duration_secs': 1.249484} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.308250] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 8fc3ea4e-d525-4583-b237-f895545fd3d1/8fc3ea4e-d525-4583-b237-f895545fd3d1.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.308250] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance '8fc3ea4e-d525-4583-b237-f895545fd3d1' progress to 50 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1090.314927] env[63355]: DEBUG oslo_concurrency.lockutils [None req-19255cf0-fd7f-4423-956e-bbebe1c80d22 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "50e7499c-a8c1-41a6-adac-36e6c3b92819" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.011s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.331143] env[63355]: DEBUG nova.policy [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bf70ebeeb3a4327ada51d18f5651476', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce0682a99ac94aeea463c961b84e6b58', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1090.569284] env[63355]: DEBUG nova.network.neutron [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Successfully created port: d1bb77fa-4894-44d2-8dba-dc7be12d4d13 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1090.795411] env[63355]: DEBUG nova.compute.manager [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1090.818294] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc6d57c-7d19-4fae-83b6-d12021adae5c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.837396] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef7cf36-e16f-4503-abfb-7a98e70b6936 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.855706] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance '8fc3ea4e-d525-4583-b237-f895545fd3d1' progress to 67 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1091.418767] env[63355]: DEBUG nova.network.neutron [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Port 6ca66441-d91f-4644-b218-fc67cf0b1ef8 binding to destination host cpu-1 is already ACTIVE {{(pid=63355) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1091.805282] env[63355]: DEBUG nova.compute.manager [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1091.832747] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.832747] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.832921] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.833942] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.833942] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.833942] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.834225] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.834518] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.834738] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.834916] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.835215] env[63355]: DEBUG nova.virt.hardware [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.836147] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d131a3e3-e02e-4859-ab36-1da9911f162b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.846566] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2da2b71-f1c7-40aa-baf4-1ce885f5c3b1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.962649] env[63355]: DEBUG nova.compute.manager [req-a3d14261-7497-4826-ab30-ec7cf4919482 req-7ab6cc2b-2534-40cc-a61f-784ac5489281 service nova] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Received event network-vif-plugged-d1bb77fa-4894-44d2-8dba-dc7be12d4d13 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1091.962816] env[63355]: DEBUG oslo_concurrency.lockutils [req-a3d14261-7497-4826-ab30-ec7cf4919482 req-7ab6cc2b-2534-40cc-a61f-784ac5489281 service nova] Acquiring lock "bcf935ad-afa2-4b36-b01b-14020453ad98-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.963051] env[63355]: DEBUG oslo_concurrency.lockutils [req-a3d14261-7497-4826-ab30-ec7cf4919482 req-7ab6cc2b-2534-40cc-a61f-784ac5489281 service nova] Lock "bcf935ad-afa2-4b36-b01b-14020453ad98-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.963233] env[63355]: DEBUG oslo_concurrency.lockutils [req-a3d14261-7497-4826-ab30-ec7cf4919482 req-7ab6cc2b-2534-40cc-a61f-784ac5489281 service nova] Lock "bcf935ad-afa2-4b36-b01b-14020453ad98-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.963408] env[63355]: DEBUG nova.compute.manager [req-a3d14261-7497-4826-ab30-ec7cf4919482 req-7ab6cc2b-2534-40cc-a61f-784ac5489281 service nova] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] No waiting events found dispatching network-vif-plugged-d1bb77fa-4894-44d2-8dba-dc7be12d4d13 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1091.963580] env[63355]: WARNING nova.compute.manager [req-a3d14261-7497-4826-ab30-ec7cf4919482 req-7ab6cc2b-2534-40cc-a61f-784ac5489281 service nova] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Received unexpected event network-vif-plugged-d1bb77fa-4894-44d2-8dba-dc7be12d4d13 for instance with vm_state building and task_state spawning. [ 1092.270807] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.270807] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.270807] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1092.270807] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Rebuilding the list of instances to heal {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1092.461515] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "8fc3ea4e-d525-4583-b237-f895545fd3d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.461515] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.461515] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.538918] env[63355]: DEBUG nova.network.neutron [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Successfully updated port: d1bb77fa-4894-44d2-8dba-dc7be12d4d13 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1092.571174] env[63355]: DEBUG nova.compute.manager [req-79241f46-c96d-4de7-91c2-5c70fe7d5ed4 req-839dc834-1e27-4c24-aaff-eb056a001a28 service nova] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Received event network-changed-d1bb77fa-4894-44d2-8dba-dc7be12d4d13 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.572866] env[63355]: DEBUG nova.compute.manager [req-79241f46-c96d-4de7-91c2-5c70fe7d5ed4 req-839dc834-1e27-4c24-aaff-eb056a001a28 service nova] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Refreshing instance network info cache due to event network-changed-d1bb77fa-4894-44d2-8dba-dc7be12d4d13. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1092.573251] env[63355]: DEBUG oslo_concurrency.lockutils [req-79241f46-c96d-4de7-91c2-5c70fe7d5ed4 req-839dc834-1e27-4c24-aaff-eb056a001a28 service nova] Acquiring lock "refresh_cache-bcf935ad-afa2-4b36-b01b-14020453ad98" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.573525] env[63355]: DEBUG oslo_concurrency.lockutils [req-79241f46-c96d-4de7-91c2-5c70fe7d5ed4 req-839dc834-1e27-4c24-aaff-eb056a001a28 service nova] Acquired lock "refresh_cache-bcf935ad-afa2-4b36-b01b-14020453ad98" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.573795] env[63355]: DEBUG nova.network.neutron [req-79241f46-c96d-4de7-91c2-5c70fe7d5ed4 req-839dc834-1e27-4c24-aaff-eb056a001a28 service nova] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Refreshing network info cache for port d1bb77fa-4894-44d2-8dba-dc7be12d4d13 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1092.775028] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Skipping network cache update for instance because it is Building. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1092.991270] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.991270] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquired lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.991270] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Forcefully refreshing network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1092.991270] env[63355]: DEBUG nova.objects.instance [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lazy-loading 'info_cache' on Instance uuid ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.040183] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "refresh_cache-bcf935ad-afa2-4b36-b01b-14020453ad98" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.114897] env[63355]: DEBUG nova.network.neutron [req-79241f46-c96d-4de7-91c2-5c70fe7d5ed4 req-839dc834-1e27-4c24-aaff-eb056a001a28 service nova] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1093.234034] env[63355]: DEBUG nova.network.neutron [req-79241f46-c96d-4de7-91c2-5c70fe7d5ed4 req-839dc834-1e27-4c24-aaff-eb056a001a28 service nova] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.505647] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.505905] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.506073] env[63355]: DEBUG nova.network.neutron [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1093.736903] env[63355]: DEBUG oslo_concurrency.lockutils [req-79241f46-c96d-4de7-91c2-5c70fe7d5ed4 req-839dc834-1e27-4c24-aaff-eb056a001a28 service nova] Releasing lock "refresh_cache-bcf935ad-afa2-4b36-b01b-14020453ad98" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.737333] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "refresh_cache-bcf935ad-afa2-4b36-b01b-14020453ad98" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.737540] env[63355]: DEBUG nova.network.neutron [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1094.232115] env[63355]: DEBUG nova.network.neutron [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance_info_cache with network_info: [{"id": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "address": "fa:16:3e:01:27:93", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca66441-d9", "ovs_interfaceid": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.271123] env[63355]: DEBUG nova.network.neutron [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1094.423585] env[63355]: DEBUG nova.network.neutron [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Updating instance_info_cache with network_info: [{"id": "d1bb77fa-4894-44d2-8dba-dc7be12d4d13", "address": "fa:16:3e:8d:22:7e", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1bb77fa-48", "ovs_interfaceid": "d1bb77fa-4894-44d2-8dba-dc7be12d4d13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.735656] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.752108] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updating instance_info_cache with network_info: [{"id": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "address": "fa:16:3e:fb:af:dc", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4c4170a-d2", "ovs_interfaceid": "d4c4170a-d27a-4f64-8789-6e01e1adb3df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.927873] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "refresh_cache-bcf935ad-afa2-4b36-b01b-14020453ad98" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.927873] env[63355]: DEBUG nova.compute.manager [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Instance network_info: |[{"id": "d1bb77fa-4894-44d2-8dba-dc7be12d4d13", "address": "fa:16:3e:8d:22:7e", "network": {"id": "cc5234cc-9dd7-42f9-ad13-c3f945583a26", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1697886957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce0682a99ac94aeea463c961b84e6b58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "130387c4-e4ec-4d95-8e9d-bb079baabad8", "external-id": "nsx-vlan-transportzone-105", "segmentation_id": 105, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1bb77fa-48", "ovs_interfaceid": "d1bb77fa-4894-44d2-8dba-dc7be12d4d13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1094.927873] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:22:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '130387c4-e4ec-4d95-8e9d-bb079baabad8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd1bb77fa-4894-44d2-8dba-dc7be12d4d13', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1094.934992] env[63355]: DEBUG oslo.service.loopingcall [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.935251] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1094.935491] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-610fe59e-b251-4c90-8c29-45b5f4e14da6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.957516] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1094.957516] env[63355]: value = "task-1350152" [ 1094.957516] env[63355]: _type = "Task" [ 1094.957516] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.966655] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350152, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.255526] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Releasing lock "refresh_cache-ae3961d2-dc5b-4e49-acca-6fb52291f23b" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.258042] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updated the network info_cache for instance {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1095.258042] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.258042] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.258042] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.258042] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.258042] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.258042] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.258042] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1095.258042] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.262675] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920aab28-bf54-4b43-a9bd-c585fb071bcb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.284078] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abf16c2-3588-42b3-b984-9200fd18c81b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.292308] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance '8fc3ea4e-d525-4583-b237-f895545fd3d1' progress to 83 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1095.465887] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350152, 'name': CreateVM_Task, 'duration_secs': 0.383466} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.466088] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1095.466782] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.466954] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.467309] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1095.467608] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a2b4e08-72e9-4d6d-8b87-273f7afebfc1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.471932] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1095.471932] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5231813c-94bf-3cec-4392-4b60bf2a5144" [ 1095.471932] env[63355]: _type = "Task" [ 1095.471932] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.479784] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5231813c-94bf-3cec-4392-4b60bf2a5144, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.761009] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.761280] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.761461] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.761634] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1095.762558] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ccb93f2-6e2e-4daa-87dc-986e6f258eb5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.769635] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.769866] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.774565] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa114dd-6ac0-430e-9250-d72ed53ad9e6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.788897] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee2352b-934c-411b-bcb8-20559132a399 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.795631] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5cda5d-93b7-4138-9917-025fe48fb667 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.801569] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1095.802019] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b36f0de-2a82-45c2-aa4c-daa9194c37b3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.831048] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180383MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1095.831218] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.831422] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.834209] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1095.834209] env[63355]: value = "task-1350153" [ 1095.834209] env[63355]: _type = "Task" [ 1095.834209] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.843288] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350153, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.982798] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5231813c-94bf-3cec-4392-4b60bf2a5144, 'name': SearchDatastore_Task, 'duration_secs': 0.010791} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.983229] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.983893] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1095.983893] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.984037] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.984140] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1095.988019] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32fee9ea-b072-46e0-8def-ef83c1ba4033 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.995165] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1095.995369] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1095.996120] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93aa737e-03fb-4b60-83e7-e63d1644b0f9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.001395] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1096.001395] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52329b12-5621-e0cb-a2a1-998d55a04d29" [ 1096.001395] env[63355]: _type = "Task" [ 1096.001395] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.009380] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52329b12-5621-e0cb-a2a1-998d55a04d29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.272618] env[63355]: DEBUG nova.compute.manager [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1096.347699] env[63355]: DEBUG oslo_vmware.api [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350153, 'name': PowerOnVM_Task, 'duration_secs': 0.387013} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.347699] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1096.347894] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b138fdd6-25bc-42ea-a920-7cb982021cb7 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance '8fc3ea4e-d525-4583-b237-f895545fd3d1' progress to 100 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1096.513770] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52329b12-5621-e0cb-a2a1-998d55a04d29, 'name': SearchDatastore_Task, 'duration_secs': 0.007719} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.514552] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e33533fa-a86e-4286-97d6-db2ce5540320 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.519700] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1096.519700] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523391dc-40c4-2f95-536d-9bd3eaa8b68c" [ 1096.519700] env[63355]: _type = "Task" [ 1096.519700] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.526776] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523391dc-40c4-2f95-536d-9bd3eaa8b68c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.793180] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.843656] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Applying migration context for instance 8fc3ea4e-d525-4583-b237-f895545fd3d1 as it has an incoming, in-progress migration a01d617f-dd26-4832-8207-3e966623a221. Migration status is post-migrating {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1096.844512] env[63355]: INFO nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating resource usage from migration a01d617f-dd26-4832-8207-3e966623a221 [ 1096.859990] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance ae3961d2-dc5b-4e49-acca-6fb52291f23b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1096.860309] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1096.861129] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Migration a01d617f-dd26-4832-8207-3e966623a221 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1096.861129] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8fc3ea4e-d525-4583-b237-f895545fd3d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1096.861129] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance bcf935ad-afa2-4b36-b01b-14020453ad98 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1097.030776] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523391dc-40c4-2f95-536d-9bd3eaa8b68c, 'name': SearchDatastore_Task, 'duration_secs': 0.010516} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.030979] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.031280] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] bcf935ad-afa2-4b36-b01b-14020453ad98/bcf935ad-afa2-4b36-b01b-14020453ad98.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1097.031539] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-846c53ae-3256-4fbd-b49c-de06f5b24a92 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.038844] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1097.038844] env[63355]: value = "task-1350154" [ 1097.038844] env[63355]: _type = "Task" [ 1097.038844] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.047214] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350154, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.364162] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1097.364162] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1097.364162] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1097.457567] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610246f8-f445-4177-aebf-5d1b490e66aa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.465432] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8446af5b-2219-4a6c-9b0a-948348b565ba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.495965] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb67f14-f1cc-41c3-b99e-0233c7f64681 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.504017] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358cbe0a-cd1b-458a-95bf-fff7e941d39b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.519697] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.549657] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350154, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.023118] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1098.049517] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350154, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568973} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.049781] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] bcf935ad-afa2-4b36-b01b-14020453ad98/bcf935ad-afa2-4b36-b01b-14020453ad98.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1098.049997] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1098.050269] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f47c546-e94d-4555-9af0-55f97218e5e1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.058109] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1098.058109] env[63355]: value = "task-1350155" [ 1098.058109] env[63355]: _type = "Task" [ 1098.058109] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.065522] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350155, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.528195] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1098.528568] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.697s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.528692] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.736s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.530722] env[63355]: INFO nova.compute.claims [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1098.567390] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350155, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062455} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.567733] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1098.568520] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb101ddc-f47b-4f07-8329-d5ced5d39107 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.590647] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] bcf935ad-afa2-4b36-b01b-14020453ad98/bcf935ad-afa2-4b36-b01b-14020453ad98.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1098.591109] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e93fc6b4-3fd1-4000-8175-5fa8b2583188 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.610618] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1098.610618] env[63355]: value = "task-1350156" [ 1098.610618] env[63355]: _type = "Task" [ 1098.610618] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.618015] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350156, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.746560] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "8fc3ea4e-d525-4583-b237-f895545fd3d1" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.746849] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.747057] env[63355]: DEBUG nova.compute.manager [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Going to confirm migration 4 {{(pid=63355) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1099.122394] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350156, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.320820] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.321047] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquired lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.321243] env[63355]: DEBUG nova.network.neutron [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1099.321435] env[63355]: DEBUG nova.objects.instance [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lazy-loading 'info_cache' on Instance uuid 8fc3ea4e-d525-4583-b237-f895545fd3d1 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.621920] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350156, 'name': ReconfigVM_Task, 'duration_secs': 0.892545} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.622240] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Reconfigured VM instance instance-00000064 to attach disk [datastore1] bcf935ad-afa2-4b36-b01b-14020453ad98/bcf935ad-afa2-4b36-b01b-14020453ad98.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.622886] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a351c1cb-1b83-4277-9ba7-e349733e41d0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.625470] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10603534-210b-4b26-a3ca-174fee24f973 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.633948] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6f167f-4b68-488e-9b58-abe7d7581fa7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.637235] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1099.637235] env[63355]: value = "task-1350157" [ 1099.637235] env[63355]: _type = "Task" [ 1099.637235] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.645905] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350157, 'name': Rename_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.669987] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24ef3cf-4938-44e5-bc7f-7602fdd14a49 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.677332] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427950a3-8c2d-4f1a-9fff-a15fe074fc6e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.690873] env[63355]: DEBUG nova.compute.provider_tree [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.147899] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350157, 'name': Rename_Task, 'duration_secs': 0.165946} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.148192] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1100.148442] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a93b5dc-9e15-4db6-8527-0a1cff7b2037 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.154967] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1100.154967] env[63355]: value = "task-1350158" [ 1100.154967] env[63355]: _type = "Task" [ 1100.154967] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.162175] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350158, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.193439] env[63355]: DEBUG nova.scheduler.client.report [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1100.524072] env[63355]: DEBUG nova.network.neutron [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance_info_cache with network_info: [{"id": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "address": "fa:16:3e:01:27:93", "network": {"id": "87cfe39d-5dec-4450-82af-cac455b8969a", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1560274740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfff1b9903264e5586119ebd3a3602de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca66441-d9", "ovs_interfaceid": "6ca66441-d91f-4644-b218-fc67cf0b1ef8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.665495] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350158, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.697892] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.169s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.698412] env[63355]: DEBUG nova.compute.manager [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1101.027269] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Releasing lock "refresh_cache-8fc3ea4e-d525-4583-b237-f895545fd3d1" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.027565] env[63355]: DEBUG nova.objects.instance [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lazy-loading 'migration_context' on Instance uuid 8fc3ea4e-d525-4583-b237-f895545fd3d1 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.166062] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350158, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.202730] env[63355]: DEBUG nova.compute.utils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1101.204399] env[63355]: DEBUG nova.compute.manager [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1101.204399] env[63355]: DEBUG nova.network.neutron [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1101.240704] env[63355]: DEBUG nova.policy [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7302e68a14984b08aef9c3af0d2a12ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '876995f179ed46b397822fa1be08ea29', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1101.490419] env[63355]: DEBUG nova.network.neutron [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Successfully created port: 2ea6bc23-1560-4552-be29-22a2d0d863ab {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1101.530823] env[63355]: DEBUG nova.objects.base [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Object Instance<8fc3ea4e-d525-4583-b237-f895545fd3d1> lazy-loaded attributes: info_cache,migration_context {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1101.531850] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99b01ad-251c-4d3c-93eb-8435a117b191 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.552604] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12e47115-8da4-4e74-aa31-cdae2410ebf6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.558827] env[63355]: DEBUG oslo_vmware.api [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1101.558827] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a16256-ac38-8ab5-fd72-62ed3a9d1980" [ 1101.558827] env[63355]: _type = "Task" [ 1101.558827] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.567920] env[63355]: DEBUG oslo_vmware.api [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a16256-ac38-8ab5-fd72-62ed3a9d1980, 'name': SearchDatastore_Task, 'duration_secs': 0.007031} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.568220] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.568456] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.667578] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350158, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.707426] env[63355]: DEBUG nova.compute.manager [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1102.166347] env[63355]: DEBUG oslo_vmware.api [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350158, 'name': PowerOnVM_Task, 'duration_secs': 1.846911} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.166432] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1102.166644] env[63355]: INFO nova.compute.manager [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Took 10.36 seconds to spawn the instance on the hypervisor. [ 1102.166846] env[63355]: DEBUG nova.compute.manager [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1102.167626] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de45d85a-fc7c-4eab-8cbf-fbb01ad03e8f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.171983] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209eaddd-f96e-49ae-b89c-c7d6a95fa7e3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.183488] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44970d0b-de2d-4529-9469-5c8ae64db349 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.216953] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e386be45-edfe-4df5-a7bd-889281fa1110 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.224792] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015cd14a-d1fd-4158-8df0-66e924c84090 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.238657] env[63355]: DEBUG nova.compute.provider_tree [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.691917] env[63355]: INFO nova.compute.manager [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Took 15.99 seconds to build instance. [ 1102.717494] env[63355]: DEBUG nova.compute.manager [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1102.741950] env[63355]: DEBUG nova.scheduler.client.report [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1102.746764] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1102.747035] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1102.747207] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1102.747395] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1102.747567] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1102.747862] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1102.747946] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1102.748125] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1102.748305] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1102.748476] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1102.748656] env[63355]: DEBUG nova.virt.hardware [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1102.749579] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a500c127-5138-4116-994b-6aa9c2180cee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.757871] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0ad821b-392c-4684-857f-c3f178f9f31e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.917403] env[63355]: DEBUG nova.compute.manager [req-34509532-54ca-4154-98dd-14c5e32e8616 req-d7932364-27fd-4576-91cb-d37c852d8cd7 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Received event network-vif-plugged-2ea6bc23-1560-4552-be29-22a2d0d863ab {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1102.917786] env[63355]: DEBUG oslo_concurrency.lockutils [req-34509532-54ca-4154-98dd-14c5e32e8616 req-d7932364-27fd-4576-91cb-d37c852d8cd7 service nova] Acquiring lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.917879] env[63355]: DEBUG oslo_concurrency.lockutils [req-34509532-54ca-4154-98dd-14c5e32e8616 req-d7932364-27fd-4576-91cb-d37c852d8cd7 service nova] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.918079] env[63355]: DEBUG oslo_concurrency.lockutils [req-34509532-54ca-4154-98dd-14c5e32e8616 req-d7932364-27fd-4576-91cb-d37c852d8cd7 service nova] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.918267] env[63355]: DEBUG nova.compute.manager [req-34509532-54ca-4154-98dd-14c5e32e8616 req-d7932364-27fd-4576-91cb-d37c852d8cd7 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] No waiting events found dispatching network-vif-plugged-2ea6bc23-1560-4552-be29-22a2d0d863ab {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1102.918438] env[63355]: WARNING nova.compute.manager [req-34509532-54ca-4154-98dd-14c5e32e8616 req-d7932364-27fd-4576-91cb-d37c852d8cd7 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Received unexpected event network-vif-plugged-2ea6bc23-1560-4552-be29-22a2d0d863ab for instance with vm_state building and task_state spawning. [ 1102.957026] env[63355]: DEBUG nova.network.neutron [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Successfully updated port: 2ea6bc23-1560-4552-be29-22a2d0d863ab {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1103.193899] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a315e55f-da2a-4738-b815-389c359f0c2e tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "bcf935ad-afa2-4b36-b01b-14020453ad98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.498s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.460134] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.460340] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.460482] env[63355]: DEBUG nova.network.neutron [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1103.758864] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.190s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.899741] env[63355]: DEBUG oslo_concurrency.lockutils [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "bcf935ad-afa2-4b36-b01b-14020453ad98" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.900066] env[63355]: DEBUG oslo_concurrency.lockutils [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "bcf935ad-afa2-4b36-b01b-14020453ad98" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.900309] env[63355]: DEBUG oslo_concurrency.lockutils [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "bcf935ad-afa2-4b36-b01b-14020453ad98-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.900501] env[63355]: DEBUG oslo_concurrency.lockutils [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "bcf935ad-afa2-4b36-b01b-14020453ad98-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.900679] env[63355]: DEBUG oslo_concurrency.lockutils [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "bcf935ad-afa2-4b36-b01b-14020453ad98-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.903157] env[63355]: INFO nova.compute.manager [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Terminating instance [ 1103.905034] env[63355]: DEBUG nova.compute.manager [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1103.905256] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1103.906130] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b837ccb-d757-4c35-a857-a08def8db786 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.913760] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1103.913993] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a45e7de4-ea05-4874-9bbe-3af254cb7b81 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.920204] env[63355]: DEBUG oslo_vmware.api [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1103.920204] env[63355]: value = "task-1350159" [ 1103.920204] env[63355]: _type = "Task" [ 1103.920204] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.929345] env[63355]: DEBUG oslo_vmware.api [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.993973] env[63355]: DEBUG nova.network.neutron [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1104.142678] env[63355]: DEBUG nova.network.neutron [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Updating instance_info_cache with network_info: [{"id": "2ea6bc23-1560-4552-be29-22a2d0d863ab", "address": "fa:16:3e:f8:7c:fc", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ea6bc23-15", "ovs_interfaceid": "2ea6bc23-1560-4552-be29-22a2d0d863ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.317563] env[63355]: INFO nova.scheduler.client.report [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted allocation for migration a01d617f-dd26-4832-8207-3e966623a221 [ 1104.431314] env[63355]: DEBUG oslo_vmware.api [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350159, 'name': PowerOffVM_Task, 'duration_secs': 0.30788} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.431591] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1104.431765] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1104.432038] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fd9e3aec-5cc0-4187-a585-8fad08728894 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.645721] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.646080] env[63355]: DEBUG nova.compute.manager [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Instance network_info: |[{"id": "2ea6bc23-1560-4552-be29-22a2d0d863ab", "address": "fa:16:3e:f8:7c:fc", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ea6bc23-15", "ovs_interfaceid": "2ea6bc23-1560-4552-be29-22a2d0d863ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1104.646516] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:7c:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ea6bc23-1560-4552-be29-22a2d0d863ab', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1104.653876] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Creating folder: Project (876995f179ed46b397822fa1be08ea29). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1104.654152] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f22f4910-88db-4cff-a161-9d33f135382f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.664111] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Created folder: Project (876995f179ed46b397822fa1be08ea29) in parent group-v287607. [ 1104.664305] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Creating folder: Instances. Parent ref: group-v287754. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1104.664531] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-56922d60-e0e5-4215-9376-dccb69cc973c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.674487] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Created folder: Instances in parent group-v287754. [ 1104.674716] env[63355]: DEBUG oslo.service.loopingcall [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1104.674913] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1104.675120] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c58ce7a-6bc8-405f-b41b-44f0f2fb54ca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.695320] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1104.695320] env[63355]: value = "task-1350163" [ 1104.695320] env[63355]: _type = "Task" [ 1104.695320] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.702705] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350163, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.823420] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.076s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.849145] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1104.849521] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1104.849831] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleting the datastore file [datastore1] bcf935ad-afa2-4b36-b01b-14020453ad98 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1104.850226] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b63f9f43-5266-4fac-aab8-81badd965f84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.858693] env[63355]: DEBUG oslo_vmware.api [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for the task: (returnval){ [ 1104.858693] env[63355]: value = "task-1350164" [ 1104.858693] env[63355]: _type = "Task" [ 1104.858693] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.870891] env[63355]: DEBUG oslo_vmware.api [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.943918] env[63355]: DEBUG nova.compute.manager [req-749f93f2-65a6-456e-a129-ed3962786a8f req-8c7b6a3b-e952-4208-b839-c7ef58b6c233 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Received event network-changed-2ea6bc23-1560-4552-be29-22a2d0d863ab {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.944276] env[63355]: DEBUG nova.compute.manager [req-749f93f2-65a6-456e-a129-ed3962786a8f req-8c7b6a3b-e952-4208-b839-c7ef58b6c233 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Refreshing instance network info cache due to event network-changed-2ea6bc23-1560-4552-be29-22a2d0d863ab. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1104.944608] env[63355]: DEBUG oslo_concurrency.lockutils [req-749f93f2-65a6-456e-a129-ed3962786a8f req-8c7b6a3b-e952-4208-b839-c7ef58b6c233 service nova] Acquiring lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.944859] env[63355]: DEBUG oslo_concurrency.lockutils [req-749f93f2-65a6-456e-a129-ed3962786a8f req-8c7b6a3b-e952-4208-b839-c7ef58b6c233 service nova] Acquired lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.945131] env[63355]: DEBUG nova.network.neutron [req-749f93f2-65a6-456e-a129-ed3962786a8f req-8c7b6a3b-e952-4208-b839-c7ef58b6c233 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Refreshing network info cache for port 2ea6bc23-1560-4552-be29-22a2d0d863ab {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1105.205010] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350163, 'name': CreateVM_Task, 'duration_secs': 0.341406} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.205196] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1105.205915] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.206111] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.206448] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1105.207038] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00c801b0-762a-4125-9e82-c1ff53efb250 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.211408] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1105.211408] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52089c16-9ab8-59ea-5272-3386059b66e9" [ 1105.211408] env[63355]: _type = "Task" [ 1105.211408] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.218758] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52089c16-9ab8-59ea-5272-3386059b66e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.368315] env[63355]: DEBUG oslo_vmware.api [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Task: {'id': task-1350164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180533} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.368572] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1105.368776] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1105.368962] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1105.369152] env[63355]: INFO nova.compute.manager [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Took 1.46 seconds to destroy the instance on the hypervisor. [ 1105.369393] env[63355]: DEBUG oslo.service.loopingcall [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1105.369584] env[63355]: DEBUG nova.compute.manager [-] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1105.369682] env[63355]: DEBUG nova.network.neutron [-] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1105.724066] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52089c16-9ab8-59ea-5272-3386059b66e9, 'name': SearchDatastore_Task, 'duration_secs': 0.009483} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.725010] env[63355]: DEBUG nova.network.neutron [req-749f93f2-65a6-456e-a129-ed3962786a8f req-8c7b6a3b-e952-4208-b839-c7ef58b6c233 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Updated VIF entry in instance network info cache for port 2ea6bc23-1560-4552-be29-22a2d0d863ab. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1105.725385] env[63355]: DEBUG nova.network.neutron [req-749f93f2-65a6-456e-a129-ed3962786a8f req-8c7b6a3b-e952-4208-b839-c7ef58b6c233 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Updating instance_info_cache with network_info: [{"id": "2ea6bc23-1560-4552-be29-22a2d0d863ab", "address": "fa:16:3e:f8:7c:fc", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ea6bc23-15", "ovs_interfaceid": "2ea6bc23-1560-4552-be29-22a2d0d863ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.726600] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.726832] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1105.727081] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.727238] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.727476] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1105.728192] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4981300-a7d3-429c-839b-eef4e8b0ded2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.735807] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1105.735975] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1105.737239] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5889bfc-0cf3-4600-97fa-9712f395c97c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.742928] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1105.742928] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5202f817-e102-5d81-08ad-20d0f74b5a56" [ 1105.742928] env[63355]: _type = "Task" [ 1105.742928] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.750180] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5202f817-e102-5d81-08ad-20d0f74b5a56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.197148] env[63355]: DEBUG nova.network.neutron [-] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.228457] env[63355]: DEBUG oslo_concurrency.lockutils [req-749f93f2-65a6-456e-a129-ed3962786a8f req-8c7b6a3b-e952-4208-b839-c7ef58b6c233 service nova] Releasing lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.254305] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5202f817-e102-5d81-08ad-20d0f74b5a56, 'name': SearchDatastore_Task, 'duration_secs': 0.007896} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.255112] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f08bb958-f4b3-4c83-abce-67bc65a03c80 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.260823] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1106.260823] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52882d7a-ee9b-3a2d-5faa-e1db770855dd" [ 1106.260823] env[63355]: _type = "Task" [ 1106.260823] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.268906] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52882d7a-ee9b-3a2d-5faa-e1db770855dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.271368] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "8fc3ea4e-d525-4583-b237-f895545fd3d1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.271600] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.271807] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "8fc3ea4e-d525-4583-b237-f895545fd3d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.272007] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.272192] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.274178] env[63355]: INFO nova.compute.manager [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Terminating instance [ 1106.277807] env[63355]: DEBUG nova.compute.manager [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1106.278032] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1106.278775] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa522d42-206c-4e0a-a923-6bbbf5936b90 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.285457] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1106.286111] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-86cc2811-6209-453e-8322-ace97de952b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.293855] env[63355]: DEBUG oslo_vmware.api [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1106.293855] env[63355]: value = "task-1350165" [ 1106.293855] env[63355]: _type = "Task" [ 1106.293855] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.301872] env[63355]: DEBUG oslo_vmware.api [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350165, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.699775] env[63355]: INFO nova.compute.manager [-] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Took 1.33 seconds to deallocate network for instance. [ 1106.771238] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52882d7a-ee9b-3a2d-5faa-e1db770855dd, 'name': SearchDatastore_Task, 'duration_secs': 0.009287} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.771516] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.771777] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd/7aee05be-b8b4-4767-b5d4-88aa4a21d5cd.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1106.772044] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a856a6b-c76e-4620-88f1-c8f87c6de492 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.778383] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1106.778383] env[63355]: value = "task-1350166" [ 1106.778383] env[63355]: _type = "Task" [ 1106.778383] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.786037] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350166, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.801638] env[63355]: DEBUG oslo_vmware.api [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350165, 'name': PowerOffVM_Task, 'duration_secs': 0.182057} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.801888] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1106.802077] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1106.802315] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5670934-ddfc-4ef7-8a19-c98bd3818413 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.861828] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1106.862073] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1106.862265] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleting the datastore file [datastore2] 8fc3ea4e-d525-4583-b237-f895545fd3d1 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1106.862533] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6be40ab7-c7f0-41bc-bcda-aee65b6650a0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.869607] env[63355]: DEBUG oslo_vmware.api [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for the task: (returnval){ [ 1106.869607] env[63355]: value = "task-1350168" [ 1106.869607] env[63355]: _type = "Task" [ 1106.869607] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.876765] env[63355]: DEBUG oslo_vmware.api [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350168, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.972479] env[63355]: DEBUG nova.compute.manager [req-9a5578ed-8c10-4814-943e-5e10e819045d req-1e548965-fae4-4837-94a3-1c37ef0ab9ab service nova] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Received event network-vif-deleted-d1bb77fa-4894-44d2-8dba-dc7be12d4d13 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1107.206687] env[63355]: DEBUG oslo_concurrency.lockutils [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.206687] env[63355]: DEBUG oslo_concurrency.lockutils [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.206687] env[63355]: DEBUG nova.objects.instance [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lazy-loading 'resources' on Instance uuid bcf935ad-afa2-4b36-b01b-14020453ad98 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.288980] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350166, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436499} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.289269] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd/7aee05be-b8b4-4767-b5d4-88aa4a21d5cd.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1107.289490] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1107.289742] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ecafce2a-dde1-4f02-99d2-c970de2328ba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.295396] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1107.295396] env[63355]: value = "task-1350169" [ 1107.295396] env[63355]: _type = "Task" [ 1107.295396] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.302461] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350169, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.378359] env[63355]: DEBUG oslo_vmware.api [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Task: {'id': task-1350168, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.393452} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.378628] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1107.378863] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1107.379109] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1107.379303] env[63355]: INFO nova.compute.manager [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1107.379548] env[63355]: DEBUG oslo.service.loopingcall [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1107.379840] env[63355]: DEBUG nova.compute.manager [-] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1107.379943] env[63355]: DEBUG nova.network.neutron [-] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1107.787086] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d16d06-764e-4a3f-916a-44e53f5acb90 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.794280] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c65295e-13f8-4ab3-a268-1f83d9debe2f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.804383] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350169, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062921} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.805098] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1107.828320] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689a66a8-a504-404f-a064-5d3cef575c94 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.831401] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d76291d-07b4-40e4-b73d-90d1ed126f32 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.846662] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abedecac-be7e-4d79-9d8c-e23ad814693f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.858258] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd/7aee05be-b8b4-4767-b5d4-88aa4a21d5cd.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1107.858483] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-389995a2-db14-4e51-868a-e19bf9157d4d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.881240] env[63355]: DEBUG nova.compute.provider_tree [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1107.883222] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1107.883222] env[63355]: value = "task-1350170" [ 1107.883222] env[63355]: _type = "Task" [ 1107.883222] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.890486] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350170, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.106994] env[63355]: DEBUG nova.network.neutron [-] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.385247] env[63355]: DEBUG nova.scheduler.client.report [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1108.396963] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350170, 'name': ReconfigVM_Task, 'duration_secs': 0.295513} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.397255] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd/7aee05be-b8b4-4767-b5d4-88aa4a21d5cd.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.397894] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc9be9d3-e61d-45a2-8cd5-bd83b525809f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.404626] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1108.404626] env[63355]: value = "task-1350171" [ 1108.404626] env[63355]: _type = "Task" [ 1108.404626] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.411833] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350171, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.610090] env[63355]: INFO nova.compute.manager [-] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Took 1.23 seconds to deallocate network for instance. [ 1108.893374] env[63355]: DEBUG oslo_concurrency.lockutils [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.687s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.913912] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350171, 'name': Rename_Task, 'duration_secs': 0.127282} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.914797] env[63355]: INFO nova.scheduler.client.report [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Deleted allocations for instance bcf935ad-afa2-4b36-b01b-14020453ad98 [ 1108.915752] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1108.918041] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4884af4d-e960-4df5-bd11-32e8a1a1b9a4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.924628] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1108.924628] env[63355]: value = "task-1350172" [ 1108.924628] env[63355]: _type = "Task" [ 1108.924628] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.932696] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350172, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.998467] env[63355]: DEBUG nova.compute.manager [req-ec2ebff3-654b-42ba-a89a-24a72f00d040 req-6e765f18-086c-4c67-84b8-f5b2dbadd26b service nova] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Received event network-vif-deleted-6ca66441-d91f-4644-b218-fc67cf0b1ef8 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1109.117296] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.117501] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.117737] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.133592] env[63355]: INFO nova.scheduler.client.report [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Deleted allocations for instance 8fc3ea4e-d525-4583-b237-f895545fd3d1 [ 1109.423968] env[63355]: DEBUG oslo_concurrency.lockutils [None req-001a217c-a9b1-403f-9966-6b4441037649 tempest-ServerDiskConfigTestJSON-1077676371 tempest-ServerDiskConfigTestJSON-1077676371-project-member] Lock "bcf935ad-afa2-4b36-b01b-14020453ad98" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.524s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.437328] env[63355]: DEBUG oslo_vmware.api [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350172, 'name': PowerOnVM_Task, 'duration_secs': 0.453283} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.438018] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1109.438154] env[63355]: INFO nova.compute.manager [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Took 6.72 seconds to spawn the instance on the hypervisor. [ 1109.438317] env[63355]: DEBUG nova.compute.manager [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1109.439432] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ce0eef-8914-4077-a121-9d0906929b3f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.641711] env[63355]: DEBUG oslo_concurrency.lockutils [None req-49e59afb-eabd-424f-ada9-7b8422d53422 tempest-DeleteServersTestJSON-373578804 tempest-DeleteServersTestJSON-373578804-project-member] Lock "8fc3ea4e-d525-4583-b237-f895545fd3d1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.370s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.958260] env[63355]: INFO nova.compute.manager [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Took 13.18 seconds to build instance. [ 1110.460465] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3e97023d-45b7-4906-8767-fddae6356aac tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.690s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.819040] env[63355]: DEBUG nova.compute.manager [req-59e8a61b-9152-4b0e-961a-0a14840cde28 req-4c748740-f57a-4569-bdb2-e60a0ed228f7 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Received event network-changed-2ea6bc23-1560-4552-be29-22a2d0d863ab {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1111.819295] env[63355]: DEBUG nova.compute.manager [req-59e8a61b-9152-4b0e-961a-0a14840cde28 req-4c748740-f57a-4569-bdb2-e60a0ed228f7 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Refreshing instance network info cache due to event network-changed-2ea6bc23-1560-4552-be29-22a2d0d863ab. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1111.819470] env[63355]: DEBUG oslo_concurrency.lockutils [req-59e8a61b-9152-4b0e-961a-0a14840cde28 req-4c748740-f57a-4569-bdb2-e60a0ed228f7 service nova] Acquiring lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.819616] env[63355]: DEBUG oslo_concurrency.lockutils [req-59e8a61b-9152-4b0e-961a-0a14840cde28 req-4c748740-f57a-4569-bdb2-e60a0ed228f7 service nova] Acquired lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.819786] env[63355]: DEBUG nova.network.neutron [req-59e8a61b-9152-4b0e-961a-0a14840cde28 req-4c748740-f57a-4569-bdb2-e60a0ed228f7 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Refreshing network info cache for port 2ea6bc23-1560-4552-be29-22a2d0d863ab {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1112.659602] env[63355]: DEBUG nova.network.neutron [req-59e8a61b-9152-4b0e-961a-0a14840cde28 req-4c748740-f57a-4569-bdb2-e60a0ed228f7 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Updated VIF entry in instance network info cache for port 2ea6bc23-1560-4552-be29-22a2d0d863ab. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1112.659801] env[63355]: DEBUG nova.network.neutron [req-59e8a61b-9152-4b0e-961a-0a14840cde28 req-4c748740-f57a-4569-bdb2-e60a0ed228f7 service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Updating instance_info_cache with network_info: [{"id": "2ea6bc23-1560-4552-be29-22a2d0d863ab", "address": "fa:16:3e:f8:7c:fc", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ea6bc23-15", "ovs_interfaceid": "2ea6bc23-1560-4552-be29-22a2d0d863ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.162921] env[63355]: DEBUG oslo_concurrency.lockutils [req-59e8a61b-9152-4b0e-961a-0a14840cde28 req-4c748740-f57a-4569-bdb2-e60a0ed228f7 service nova] Releasing lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.695814] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.695814] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.695814] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.695814] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.695814] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.696656] env[63355]: INFO nova.compute.manager [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Terminating instance [ 1113.698574] env[63355]: DEBUG nova.compute.manager [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1113.698908] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1113.699860] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb569e59-9871-44d5-8ee4-5d2671828940 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.708127] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1113.708522] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40df18a9-e6c8-4c41-a822-c2089521d15e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.715408] env[63355]: DEBUG oslo_vmware.api [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1113.715408] env[63355]: value = "task-1350173" [ 1113.715408] env[63355]: _type = "Task" [ 1113.715408] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.723980] env[63355]: DEBUG oslo_vmware.api [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350173, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.226511] env[63355]: DEBUG oslo_vmware.api [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350173, 'name': PowerOffVM_Task, 'duration_secs': 0.233346} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.227172] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1114.227521] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1114.227899] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14ce592e-6476-4fd3-bd3a-3da8ec205212 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.297021] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1114.297021] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1114.297021] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Deleting the datastore file [datastore2] ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1114.297021] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-737940c0-a9c9-4d4a-9c5c-f15b93982783 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.302426] env[63355]: DEBUG oslo_vmware.api [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1114.302426] env[63355]: value = "task-1350175" [ 1114.302426] env[63355]: _type = "Task" [ 1114.302426] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.309826] env[63355]: DEBUG oslo_vmware.api [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350175, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.812526] env[63355]: DEBUG oslo_vmware.api [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350175, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.413594} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.815931] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1114.815931] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1114.815931] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1114.815931] env[63355]: INFO nova.compute.manager [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1114.815931] env[63355]: DEBUG oslo.service.loopingcall [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1114.815931] env[63355]: DEBUG nova.compute.manager [-] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1114.815931] env[63355]: DEBUG nova.network.neutron [-] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1115.630361] env[63355]: DEBUG nova.compute.manager [req-c01ab2b9-e1b4-4c32-aefd-6f0cc470daf3 req-59ccaa99-4489-452e-a7ad-7c8c2ef22a29 service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Received event network-vif-deleted-d4c4170a-d27a-4f64-8789-6e01e1adb3df {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.630775] env[63355]: INFO nova.compute.manager [req-c01ab2b9-e1b4-4c32-aefd-6f0cc470daf3 req-59ccaa99-4489-452e-a7ad-7c8c2ef22a29 service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Neutron deleted interface d4c4170a-d27a-4f64-8789-6e01e1adb3df; detaching it from the instance and deleting it from the info cache [ 1115.630775] env[63355]: DEBUG nova.network.neutron [req-c01ab2b9-e1b4-4c32-aefd-6f0cc470daf3 req-59ccaa99-4489-452e-a7ad-7c8c2ef22a29 service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.094530] env[63355]: DEBUG nova.network.neutron [-] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.132775] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f443744a-4630-47c1-920c-b5c7f0b228b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.142999] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b396747-db84-4861-a9f2-e344089ae5e6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.170308] env[63355]: DEBUG nova.compute.manager [req-c01ab2b9-e1b4-4c32-aefd-6f0cc470daf3 req-59ccaa99-4489-452e-a7ad-7c8c2ef22a29 service nova] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Detach interface failed, port_id=d4c4170a-d27a-4f64-8789-6e01e1adb3df, reason: Instance ae3961d2-dc5b-4e49-acca-6fb52291f23b could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1116.596856] env[63355]: INFO nova.compute.manager [-] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Took 1.78 seconds to deallocate network for instance. [ 1117.105162] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.105590] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.105769] env[63355]: DEBUG nova.objects.instance [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'resources' on Instance uuid ae3961d2-dc5b-4e49-acca-6fb52291f23b {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.680349] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf28600-4e25-42c7-9a76-98265340333d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.689657] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc547fef-a2cd-4d4b-9e80-5284e36e065e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.726750] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7635f95d-1096-4a09-98f4-b42325bf5a3e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.734657] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03350485-60e1-4c11-902e-278a4f1bce77 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.749503] env[63355]: DEBUG nova.compute.provider_tree [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.879246] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "8db23054-4556-496a-a4b1-a24c71c30f50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.879518] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "8db23054-4556-496a-a4b1-a24c71c30f50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.252529] env[63355]: DEBUG nova.scheduler.client.report [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.385999] env[63355]: DEBUG nova.compute.manager [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1118.760299] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.654s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.780736] env[63355]: INFO nova.scheduler.client.report [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Deleted allocations for instance ae3961d2-dc5b-4e49-acca-6fb52291f23b [ 1118.906023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.906023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.907482] env[63355]: INFO nova.compute.claims [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.181045] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquiring lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.181560] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.290815] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4d23c8f4-9091-41dd-9a8a-2aa83cace34f tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "ae3961d2-dc5b-4e49-acca-6fb52291f23b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.597s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.686064] env[63355]: DEBUG nova.compute.manager [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1119.991843] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985baec5-3c76-414e-b958-453c043721fa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.007929] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283f12b3-3aab-40cb-8fe3-8a5c4ce12a24 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.039892] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2442dde9-781e-4e73-a9a2-a5cc271463cf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.047299] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee42d58f-d8b2-4b4a-b6a6-fb8fa11d1f09 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.060383] env[63355]: DEBUG nova.compute.provider_tree [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.207653] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.563384] env[63355]: DEBUG nova.scheduler.client.report [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.072120] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.164s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.072120] env[63355]: DEBUG nova.compute.manager [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1121.074112] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.867s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.078168] env[63355]: INFO nova.compute.claims [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1121.302811] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.303186] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.586584] env[63355]: DEBUG nova.compute.utils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1121.588104] env[63355]: DEBUG nova.compute.manager [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1121.588313] env[63355]: DEBUG nova.network.neutron [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1121.633345] env[63355]: DEBUG nova.policy [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a61bc098ea4e43ceab263ae9890da7c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff5b796e7bb0458dbd3b862dcae7bdd1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1121.806317] env[63355]: DEBUG nova.compute.utils [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1121.829235] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.829482] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.916284] env[63355]: DEBUG nova.network.neutron [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Successfully created port: e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1122.093997] env[63355]: DEBUG nova.compute.manager [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1122.177045] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090745ba-caa4-4fe3-9cf4-c0469c15a5e9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.185619] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b9d49e-fb94-4acc-a519-f65fd47d1672 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.215197] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7fcb22c-16df-468e-955c-136fe54a231a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.222905] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1755cdfb-493e-4121-816a-dad6139080e2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.236184] env[63355]: DEBUG nova.compute.provider_tree [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.309837] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.331620] env[63355]: DEBUG nova.compute.manager [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1122.739839] env[63355]: DEBUG nova.scheduler.client.report [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1122.851029] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.104934] env[63355]: DEBUG nova.compute.manager [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1123.130498] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1123.130775] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1123.130945] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.131751] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1123.131751] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.131751] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1123.131751] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1123.131912] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1123.132028] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1123.132145] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1123.132324] env[63355]: DEBUG nova.virt.hardware [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.133213] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a2f4f1-af67-40cb-b676-8921c8b9ecee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.142263] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee576a5d-30ad-4d70-820f-3abde43adc89 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.244893] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.245505] env[63355]: DEBUG nova.compute.manager [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1123.248389] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.397s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.249987] env[63355]: INFO nova.compute.claims [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1123.352165] env[63355]: DEBUG nova.compute.manager [req-f4e1be95-28ca-49a2-9715-606cc7904d35 req-da47bc89-a7b4-450b-9e90-adf2a76f165a service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Received event network-vif-plugged-e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1123.352402] env[63355]: DEBUG oslo_concurrency.lockutils [req-f4e1be95-28ca-49a2-9715-606cc7904d35 req-da47bc89-a7b4-450b-9e90-adf2a76f165a service nova] Acquiring lock "8db23054-4556-496a-a4b1-a24c71c30f50-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.352614] env[63355]: DEBUG oslo_concurrency.lockutils [req-f4e1be95-28ca-49a2-9715-606cc7904d35 req-da47bc89-a7b4-450b-9e90-adf2a76f165a service nova] Lock "8db23054-4556-496a-a4b1-a24c71c30f50-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.352786] env[63355]: DEBUG oslo_concurrency.lockutils [req-f4e1be95-28ca-49a2-9715-606cc7904d35 req-da47bc89-a7b4-450b-9e90-adf2a76f165a service nova] Lock "8db23054-4556-496a-a4b1-a24c71c30f50-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.352958] env[63355]: DEBUG nova.compute.manager [req-f4e1be95-28ca-49a2-9715-606cc7904d35 req-da47bc89-a7b4-450b-9e90-adf2a76f165a service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] No waiting events found dispatching network-vif-plugged-e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1123.353312] env[63355]: WARNING nova.compute.manager [req-f4e1be95-28ca-49a2-9715-606cc7904d35 req-da47bc89-a7b4-450b-9e90-adf2a76f165a service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Received unexpected event network-vif-plugged-e3455ba6-ed91-4901-b232-76ca673421a7 for instance with vm_state building and task_state spawning. [ 1123.380591] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.380848] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.381186] env[63355]: INFO nova.compute.manager [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Attaching volume 2540a76c-67a2-4e5f-853c-f70f7cef4420 to /dev/sdb [ 1123.416500] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255a4656-9c57-4467-92c4-d10fc26e3f76 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.423649] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2daa72-7a4a-4010-baa3-9080f719f7f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.434687] env[63355]: DEBUG nova.network.neutron [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Successfully updated port: e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1123.438282] env[63355]: DEBUG nova.virt.block_device [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating existing volume attachment record: 2ecf59df-2e3b-402b-8f2a-e430a364fd49 {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1123.754197] env[63355]: DEBUG nova.compute.utils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1123.757605] env[63355]: DEBUG nova.compute.manager [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1123.757777] env[63355]: DEBUG nova.network.neutron [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1123.796225] env[63355]: DEBUG nova.policy [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6bdf3f0b33e54366aee7e92fc8eac432', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e8b0ab453c947deb7329857546b817b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1123.941746] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.941931] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.942160] env[63355]: DEBUG nova.network.neutron [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1124.029591] env[63355]: DEBUG nova.network.neutron [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Successfully created port: 8c03f4b6-4e4b-4300-932c-1c26be9d108f {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1124.263518] env[63355]: DEBUG nova.compute.manager [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1124.345024] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be17176e-90fd-4404-afe0-63febb9636f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.352597] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7782ab8b-e8d8-4a36-b3a9-106049212375 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.384310] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63409599-05b9-40d1-959e-aa73a29b40f6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.392142] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58047dea-1f4f-4bd0-acf6-ed2c9668dd3d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.405276] env[63355]: DEBUG nova.compute.provider_tree [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.472651] env[63355]: DEBUG nova.network.neutron [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1124.591387] env[63355]: DEBUG nova.network.neutron [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updating instance_info_cache with network_info: [{"id": "e3455ba6-ed91-4901-b232-76ca673421a7", "address": "fa:16:3e:eb:3b:b7", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3455ba6-ed", "ovs_interfaceid": "e3455ba6-ed91-4901-b232-76ca673421a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.908468] env[63355]: DEBUG nova.scheduler.client.report [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1125.094060] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.094424] env[63355]: DEBUG nova.compute.manager [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Instance network_info: |[{"id": "e3455ba6-ed91-4901-b232-76ca673421a7", "address": "fa:16:3e:eb:3b:b7", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3455ba6-ed", "ovs_interfaceid": "e3455ba6-ed91-4901-b232-76ca673421a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1125.094858] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:3b:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e3455ba6-ed91-4901-b232-76ca673421a7', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1125.102305] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Creating folder: Project (ff5b796e7bb0458dbd3b862dcae7bdd1). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1125.102586] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a8c5cf08-f048-40ad-a396-b8a49481f677 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.114466] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Created folder: Project (ff5b796e7bb0458dbd3b862dcae7bdd1) in parent group-v287607. [ 1125.114652] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Creating folder: Instances. Parent ref: group-v287759. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1125.114875] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4654ab4-6d75-4f9a-a2c4-da60a491215b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.124686] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Created folder: Instances in parent group-v287759. [ 1125.124955] env[63355]: DEBUG oslo.service.loopingcall [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1125.125179] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1125.125384] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95a38225-dd22-412a-b646-9177fed08fca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.144387] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1125.144387] env[63355]: value = "task-1350181" [ 1125.144387] env[63355]: _type = "Task" [ 1125.144387] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.153225] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350181, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.273189] env[63355]: DEBUG nova.compute.manager [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1125.300809] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1125.301086] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1125.301253] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1125.301439] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1125.301589] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1125.301741] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1125.301953] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1125.302130] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1125.302302] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1125.302470] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1125.302645] env[63355]: DEBUG nova.virt.hardware [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1125.303508] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459c1174-5708-4cf6-aebe-e68a2978a46a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.311257] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f054677-3615-48ac-908e-01f146e612a6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.382316] env[63355]: DEBUG nova.compute.manager [req-65925786-3a7e-44c7-8b24-518a9c4db475 req-9978c090-a10e-4e00-8309-651bb5ec9ded service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Received event network-changed-e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1125.382516] env[63355]: DEBUG nova.compute.manager [req-65925786-3a7e-44c7-8b24-518a9c4db475 req-9978c090-a10e-4e00-8309-651bb5ec9ded service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Refreshing instance network info cache due to event network-changed-e3455ba6-ed91-4901-b232-76ca673421a7. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1125.383301] env[63355]: DEBUG oslo_concurrency.lockutils [req-65925786-3a7e-44c7-8b24-518a9c4db475 req-9978c090-a10e-4e00-8309-651bb5ec9ded service nova] Acquiring lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.383301] env[63355]: DEBUG oslo_concurrency.lockutils [req-65925786-3a7e-44c7-8b24-518a9c4db475 req-9978c090-a10e-4e00-8309-651bb5ec9ded service nova] Acquired lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.383301] env[63355]: DEBUG nova.network.neutron [req-65925786-3a7e-44c7-8b24-518a9c4db475 req-9978c090-a10e-4e00-8309-651bb5ec9ded service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Refreshing network info cache for port e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1125.414187] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.166s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.414989] env[63355]: DEBUG nova.compute.manager [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1125.492870] env[63355]: DEBUG nova.network.neutron [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Successfully updated port: 8c03f4b6-4e4b-4300-932c-1c26be9d108f {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1125.653914] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350181, 'name': CreateVM_Task, 'duration_secs': 0.313356} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.654096] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1125.654778] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.654953] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.655301] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1125.655560] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-445ef39e-b6ec-4cf8-9f88-219729a41040 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.659683] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1125.659683] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d1990c-09fc-4e26-fdea-95785888202d" [ 1125.659683] env[63355]: _type = "Task" [ 1125.659683] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.666768] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d1990c-09fc-4e26-fdea-95785888202d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.921933] env[63355]: DEBUG nova.compute.utils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1125.923374] env[63355]: DEBUG nova.compute.manager [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1125.923875] env[63355]: DEBUG nova.network.neutron [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1125.992803] env[63355]: DEBUG nova.policy [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d48712ba84642d1aab85b9976659a5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58d215a4ff55488f931814352915d256', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1125.995731] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquiring lock "refresh_cache-4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.995856] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquired lock "refresh_cache-4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.996012] env[63355]: DEBUG nova.network.neutron [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1126.170532] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d1990c-09fc-4e26-fdea-95785888202d, 'name': SearchDatastore_Task, 'duration_secs': 0.009407} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.170884] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.171086] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1126.171336] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.171488] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.171673] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1126.171962] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56e86a79-8150-466e-944a-f949a291279a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.179972] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1126.180284] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1126.181040] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed203360-8ef1-43cd-a6db-c55b0bbc6e4a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.186384] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1126.186384] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522bb5d9-58bf-0f5f-fd7c-5b33ba7898dc" [ 1126.186384] env[63355]: _type = "Task" [ 1126.186384] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.193960] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522bb5d9-58bf-0f5f-fd7c-5b33ba7898dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.196716] env[63355]: DEBUG nova.network.neutron [req-65925786-3a7e-44c7-8b24-518a9c4db475 req-9978c090-a10e-4e00-8309-651bb5ec9ded service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updated VIF entry in instance network info cache for port e3455ba6-ed91-4901-b232-76ca673421a7. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1126.197050] env[63355]: DEBUG nova.network.neutron [req-65925786-3a7e-44c7-8b24-518a9c4db475 req-9978c090-a10e-4e00-8309-651bb5ec9ded service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updating instance_info_cache with network_info: [{"id": "e3455ba6-ed91-4901-b232-76ca673421a7", "address": "fa:16:3e:eb:3b:b7", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3455ba6-ed", "ovs_interfaceid": "e3455ba6-ed91-4901-b232-76ca673421a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.375543] env[63355]: DEBUG nova.network.neutron [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Successfully created port: 69950442-ccce-4a0a-95ad-cc51720ad0a6 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1126.426520] env[63355]: DEBUG nova.compute.manager [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1126.546212] env[63355]: DEBUG nova.network.neutron [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1126.697778] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522bb5d9-58bf-0f5f-fd7c-5b33ba7898dc, 'name': SearchDatastore_Task, 'duration_secs': 0.008741} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.698598] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-445eb1ea-6145-472b-b66e-839c1f9b4b89 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.700918] env[63355]: DEBUG oslo_concurrency.lockutils [req-65925786-3a7e-44c7-8b24-518a9c4db475 req-9978c090-a10e-4e00-8309-651bb5ec9ded service nova] Releasing lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.704019] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1126.704019] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5264ddd7-a11c-7fe7-8618-26da64d93d27" [ 1126.704019] env[63355]: _type = "Task" [ 1126.704019] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.704831] env[63355]: DEBUG nova.network.neutron [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Updating instance_info_cache with network_info: [{"id": "8c03f4b6-4e4b-4300-932c-1c26be9d108f", "address": "fa:16:3e:05:7a:95", "network": {"id": "fedaec57-d50a-4898-8e24-42bf7922ac0c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1831387889-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e8b0ab453c947deb7329857546b817b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c03f4b6-4e", "ovs_interfaceid": "8c03f4b6-4e4b-4300-932c-1c26be9d108f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.712755] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5264ddd7-a11c-7fe7-8618-26da64d93d27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.210354] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Releasing lock "refresh_cache-4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.210673] env[63355]: DEBUG nova.compute.manager [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Instance network_info: |[{"id": "8c03f4b6-4e4b-4300-932c-1c26be9d108f", "address": "fa:16:3e:05:7a:95", "network": {"id": "fedaec57-d50a-4898-8e24-42bf7922ac0c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1831387889-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e8b0ab453c947deb7329857546b817b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c03f4b6-4e", "ovs_interfaceid": "8c03f4b6-4e4b-4300-932c-1c26be9d108f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1127.211106] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:7a:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c03f4b6-4e4b-4300-932c-1c26be9d108f', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1127.218580] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Creating folder: Project (9e8b0ab453c947deb7329857546b817b). Parent ref: group-v287607. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1127.219252] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-84ec798b-ec53-4806-a387-16c2f7d7e02c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.224637] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5264ddd7-a11c-7fe7-8618-26da64d93d27, 'name': SearchDatastore_Task, 'duration_secs': 0.009654} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.224872] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.225137] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 8db23054-4556-496a-a4b1-a24c71c30f50/8db23054-4556-496a-a4b1-a24c71c30f50.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1127.225389] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2eb86971-03d2-41f8-88d6-5909f655623b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.229338] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Created folder: Project (9e8b0ab453c947deb7329857546b817b) in parent group-v287607. [ 1127.229519] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Creating folder: Instances. Parent ref: group-v287762. {{(pid=63355) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1127.229719] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f1d1dec-ad9e-46a7-8e43-df8cfb4c2863 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.231922] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1127.231922] env[63355]: value = "task-1350184" [ 1127.231922] env[63355]: _type = "Task" [ 1127.231922] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.239711] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.240757] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Created folder: Instances in parent group-v287762. [ 1127.240995] env[63355]: DEBUG oslo.service.loopingcall [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.241183] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1127.241375] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82db34d2-04bb-45d8-8ef9-b752586d2ba1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.260018] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1127.260018] env[63355]: value = "task-1350186" [ 1127.260018] env[63355]: _type = "Task" [ 1127.260018] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.269547] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350186, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.411789] env[63355]: DEBUG nova.compute.manager [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Received event network-vif-plugged-8c03f4b6-4e4b-4300-932c-1c26be9d108f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1127.412034] env[63355]: DEBUG oslo_concurrency.lockutils [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] Acquiring lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.412275] env[63355]: DEBUG oslo_concurrency.lockutils [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] Lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.412448] env[63355]: DEBUG oslo_concurrency.lockutils [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] Lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.412639] env[63355]: DEBUG nova.compute.manager [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] No waiting events found dispatching network-vif-plugged-8c03f4b6-4e4b-4300-932c-1c26be9d108f {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1127.412811] env[63355]: WARNING nova.compute.manager [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Received unexpected event network-vif-plugged-8c03f4b6-4e4b-4300-932c-1c26be9d108f for instance with vm_state building and task_state spawning. [ 1127.412975] env[63355]: DEBUG nova.compute.manager [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Received event network-changed-8c03f4b6-4e4b-4300-932c-1c26be9d108f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1127.413153] env[63355]: DEBUG nova.compute.manager [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Refreshing instance network info cache due to event network-changed-8c03f4b6-4e4b-4300-932c-1c26be9d108f. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1127.413347] env[63355]: DEBUG oslo_concurrency.lockutils [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] Acquiring lock "refresh_cache-4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.413485] env[63355]: DEBUG oslo_concurrency.lockutils [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] Acquired lock "refresh_cache-4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.413644] env[63355]: DEBUG nova.network.neutron [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Refreshing network info cache for port 8c03f4b6-4e4b-4300-932c-1c26be9d108f {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1127.437438] env[63355]: DEBUG nova.compute.manager [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1127.463815] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1127.464098] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1127.464271] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1127.464460] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1127.464615] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1127.464769] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1127.464987] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1127.465168] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1127.465342] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1127.465508] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1127.465686] env[63355]: DEBUG nova.virt.hardware [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1127.466616] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76934d23-7f58-4c04-bc0a-630a8724d76a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.477090] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52d69dc-e533-4fc0-8737-ef5a8d8b4189 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.744625] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500761} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.744625] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 8db23054-4556-496a-a4b1-a24c71c30f50/8db23054-4556-496a-a4b1-a24c71c30f50.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1127.744625] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1127.744625] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69436869-3435-40e6-b308-f22b7cb98175 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.751832] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1127.751832] env[63355]: value = "task-1350187" [ 1127.751832] env[63355]: _type = "Task" [ 1127.751832] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.761536] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350187, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.769686] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350186, 'name': CreateVM_Task, 'duration_secs': 0.497576} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.769853] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1127.770564] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.770743] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.771099] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1127.771380] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1861768-142e-474f-9729-3b45670cb8de {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.775649] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for the task: (returnval){ [ 1127.775649] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527b512a-b380-e69c-05cb-93c7e3e6e8fc" [ 1127.775649] env[63355]: _type = "Task" [ 1127.775649] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.783197] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527b512a-b380-e69c-05cb-93c7e3e6e8fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.906580] env[63355]: DEBUG nova.network.neutron [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Successfully updated port: 69950442-ccce-4a0a-95ad-cc51720ad0a6 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1127.988349] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1127.988551] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287758', 'volume_id': '2540a76c-67a2-4e5f-853c-f70f7cef4420', 'name': 'volume-2540a76c-67a2-4e5f-853c-f70f7cef4420', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50', 'attached_at': '', 'detached_at': '', 'volume_id': '2540a76c-67a2-4e5f-853c-f70f7cef4420', 'serial': '2540a76c-67a2-4e5f-853c-f70f7cef4420'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1127.989552] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6982aad-70ff-4d7c-8faf-92129b56e8d9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.007983] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed86a028-e148-4e52-a2c5-5614b7556488 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.032365] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] volume-2540a76c-67a2-4e5f-853c-f70f7cef4420/volume-2540a76c-67a2-4e5f-853c-f70f7cef4420.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.034547] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2162f07-8976-4672-9c2f-c246dda37490 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.051565] env[63355]: DEBUG oslo_vmware.api [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1128.051565] env[63355]: value = "task-1350188" [ 1128.051565] env[63355]: _type = "Task" [ 1128.051565] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.061918] env[63355]: DEBUG oslo_vmware.api [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350188, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.155120] env[63355]: DEBUG nova.network.neutron [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Updated VIF entry in instance network info cache for port 8c03f4b6-4e4b-4300-932c-1c26be9d108f. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1128.155545] env[63355]: DEBUG nova.network.neutron [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Updating instance_info_cache with network_info: [{"id": "8c03f4b6-4e4b-4300-932c-1c26be9d108f", "address": "fa:16:3e:05:7a:95", "network": {"id": "fedaec57-d50a-4898-8e24-42bf7922ac0c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1831387889-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e8b0ab453c947deb7329857546b817b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c03f4b6-4e", "ovs_interfaceid": "8c03f4b6-4e4b-4300-932c-1c26be9d108f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.262695] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350187, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063329} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.262695] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1128.266020] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb87cbb-9615-47ce-aca8-6138163c41a5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.285829] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 8db23054-4556-496a-a4b1-a24c71c30f50/8db23054-4556-496a-a4b1-a24c71c30f50.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.288785] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a338c278-279d-481f-b276-25da09d25bb2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.308996] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527b512a-b380-e69c-05cb-93c7e3e6e8fc, 'name': SearchDatastore_Task, 'duration_secs': 0.009735} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.310219] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.310527] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1128.310834] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.311082] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.311348] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.311704] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1128.311704] env[63355]: value = "task-1350189" [ 1128.311704] env[63355]: _type = "Task" [ 1128.311704] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.311990] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-08715427-b810-4ede-afb1-198383fb2672 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.321328] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350189, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.323172] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.323419] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1128.324125] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a86cdd4-c00d-4d6e-8373-38e563979e04 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.328616] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for the task: (returnval){ [ 1128.328616] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f11c8f-5dd6-e4b2-fc34-316062390acc" [ 1128.328616] env[63355]: _type = "Task" [ 1128.328616] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.337707] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f11c8f-5dd6-e4b2-fc34-316062390acc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.409711] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "refresh_cache-8d332e8f-4c89-4050-b5b7-3cd034f9edfe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.409933] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquired lock "refresh_cache-8d332e8f-4c89-4050-b5b7-3cd034f9edfe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.410202] env[63355]: DEBUG nova.network.neutron [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1128.561099] env[63355]: DEBUG oslo_vmware.api [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350188, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.658110] env[63355]: DEBUG oslo_concurrency.lockutils [req-575bac94-7c58-46bc-98f6-166cc23db1d4 req-7e439a63-98dd-43e0-8e41-2ecd48a65c3d service nova] Releasing lock "refresh_cache-4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.823360] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350189, 'name': ReconfigVM_Task, 'duration_secs': 0.289748} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.823624] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 8db23054-4556-496a-a4b1-a24c71c30f50/8db23054-4556-496a-a4b1-a24c71c30f50.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.824274] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec446783-57ae-4d60-81ca-96e8158fec8c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.833198] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1128.833198] env[63355]: value = "task-1350190" [ 1128.833198] env[63355]: _type = "Task" [ 1128.833198] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.839621] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52f11c8f-5dd6-e4b2-fc34-316062390acc, 'name': SearchDatastore_Task, 'duration_secs': 0.010574} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.840620] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf31df31-6698-4c9f-bda6-3704c3a242bf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.845593] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350190, 'name': Rename_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.848256] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for the task: (returnval){ [ 1128.848256] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523f7b62-8df0-7e14-3266-4203a201368d" [ 1128.848256] env[63355]: _type = "Task" [ 1128.848256] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.855739] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523f7b62-8df0-7e14-3266-4203a201368d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.942264] env[63355]: DEBUG nova.network.neutron [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1129.061492] env[63355]: DEBUG oslo_vmware.api [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350188, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.065640] env[63355]: DEBUG nova.network.neutron [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Updating instance_info_cache with network_info: [{"id": "69950442-ccce-4a0a-95ad-cc51720ad0a6", "address": "fa:16:3e:97:4f:85", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69950442-cc", "ovs_interfaceid": "69950442-ccce-4a0a-95ad-cc51720ad0a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.343916] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350190, 'name': Rename_Task, 'duration_secs': 0.144446} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.344238] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1129.344493] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1c32c4e-4c67-4b43-85ab-2e8508de1205 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.353190] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1129.353190] env[63355]: value = "task-1350191" [ 1129.353190] env[63355]: _type = "Task" [ 1129.353190] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.359150] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523f7b62-8df0-7e14-3266-4203a201368d, 'name': SearchDatastore_Task, 'duration_secs': 0.009587} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.359697] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.359971] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe/4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1129.360230] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae7bf6e4-5015-496d-ad94-2f0055cfaca9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.364530] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350191, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.368309] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for the task: (returnval){ [ 1129.368309] env[63355]: value = "task-1350192" [ 1129.368309] env[63355]: _type = "Task" [ 1129.368309] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.375496] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350192, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.440094] env[63355]: DEBUG nova.compute.manager [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Received event network-vif-plugged-69950442-ccce-4a0a-95ad-cc51720ad0a6 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.440386] env[63355]: DEBUG oslo_concurrency.lockutils [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.440621] env[63355]: DEBUG oslo_concurrency.lockutils [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.440680] env[63355]: DEBUG oslo_concurrency.lockutils [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.440850] env[63355]: DEBUG nova.compute.manager [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] No waiting events found dispatching network-vif-plugged-69950442-ccce-4a0a-95ad-cc51720ad0a6 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1129.441135] env[63355]: WARNING nova.compute.manager [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Received unexpected event network-vif-plugged-69950442-ccce-4a0a-95ad-cc51720ad0a6 for instance with vm_state building and task_state spawning. [ 1129.441412] env[63355]: DEBUG nova.compute.manager [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Received event network-changed-69950442-ccce-4a0a-95ad-cc51720ad0a6 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.441539] env[63355]: DEBUG nova.compute.manager [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Refreshing instance network info cache due to event network-changed-69950442-ccce-4a0a-95ad-cc51720ad0a6. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1129.441696] env[63355]: DEBUG oslo_concurrency.lockutils [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] Acquiring lock "refresh_cache-8d332e8f-4c89-4050-b5b7-3cd034f9edfe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.564717] env[63355]: DEBUG oslo_vmware.api [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350188, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.568701] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Releasing lock "refresh_cache-8d332e8f-4c89-4050-b5b7-3cd034f9edfe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.569250] env[63355]: DEBUG nova.compute.manager [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Instance network_info: |[{"id": "69950442-ccce-4a0a-95ad-cc51720ad0a6", "address": "fa:16:3e:97:4f:85", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69950442-cc", "ovs_interfaceid": "69950442-ccce-4a0a-95ad-cc51720ad0a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1129.569670] env[63355]: DEBUG oslo_concurrency.lockutils [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] Acquired lock "refresh_cache-8d332e8f-4c89-4050-b5b7-3cd034f9edfe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.569910] env[63355]: DEBUG nova.network.neutron [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Refreshing network info cache for port 69950442-ccce-4a0a-95ad-cc51720ad0a6 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1129.571474] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:4f:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd986680e-ad16-45b1-bf6d-cd2fe661679f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69950442-ccce-4a0a-95ad-cc51720ad0a6', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1129.579752] env[63355]: DEBUG oslo.service.loopingcall [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1129.581184] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1129.581483] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-425ce36e-89ca-4b9b-b309-6e64fc003471 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.602572] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1129.602572] env[63355]: value = "task-1350193" [ 1129.602572] env[63355]: _type = "Task" [ 1129.602572] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.613009] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350193, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.863598] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350191, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.877044] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350192, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461862} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.877348] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe/4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1129.877574] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1129.877832] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e18a244-2a2b-498d-a699-c6f2359ef4f4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.884419] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for the task: (returnval){ [ 1129.884419] env[63355]: value = "task-1350194" [ 1129.884419] env[63355]: _type = "Task" [ 1129.884419] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.892999] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350194, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.063739] env[63355]: DEBUG oslo_vmware.api [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350188, 'name': ReconfigVM_Task, 'duration_secs': 1.707644} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.063972] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfigured VM instance instance-00000063 to attach disk [datastore2] volume-2540a76c-67a2-4e5f-853c-f70f7cef4420/volume-2540a76c-67a2-4e5f-853c-f70f7cef4420.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1130.068624] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-532b563f-a093-4659-b1bd-967706ac3259 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.083526] env[63355]: DEBUG oslo_vmware.api [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1130.083526] env[63355]: value = "task-1350195" [ 1130.083526] env[63355]: _type = "Task" [ 1130.083526] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.091406] env[63355]: DEBUG oslo_vmware.api [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350195, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.111720] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350193, 'name': CreateVM_Task, 'duration_secs': 0.432374} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.111932] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1130.112680] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.112885] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.113253] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1130.115654] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fea5c4b-8639-46ba-80ce-b73651a11eb5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.120805] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1130.120805] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fd2792-6a8d-4510-9035-4e67a7ce8dae" [ 1130.120805] env[63355]: _type = "Task" [ 1130.120805] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.128513] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fd2792-6a8d-4510-9035-4e67a7ce8dae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.364344] env[63355]: DEBUG oslo_vmware.api [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350191, 'name': PowerOnVM_Task, 'duration_secs': 0.633845} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.364634] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1130.364836] env[63355]: INFO nova.compute.manager [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Took 7.26 seconds to spawn the instance on the hypervisor. [ 1130.365028] env[63355]: DEBUG nova.compute.manager [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1130.365792] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7175a1c4-5b2c-4305-88ea-4bf23db354d9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.392950] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350194, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069775} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.393231] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1130.393982] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b027494-0975-45cb-93e6-bb48aeeb8d99 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.415304] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe/4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1130.417696] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b110a0b-eeec-49f2-baf8-c1cef6fc45f5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.438561] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for the task: (returnval){ [ 1130.438561] env[63355]: value = "task-1350196" [ 1130.438561] env[63355]: _type = "Task" [ 1130.438561] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.446271] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350196, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.471560] env[63355]: DEBUG nova.network.neutron [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Updated VIF entry in instance network info cache for port 69950442-ccce-4a0a-95ad-cc51720ad0a6. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1130.471784] env[63355]: DEBUG nova.network.neutron [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Updating instance_info_cache with network_info: [{"id": "69950442-ccce-4a0a-95ad-cc51720ad0a6", "address": "fa:16:3e:97:4f:85", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69950442-cc", "ovs_interfaceid": "69950442-ccce-4a0a-95ad-cc51720ad0a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.593397] env[63355]: DEBUG oslo_vmware.api [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350195, 'name': ReconfigVM_Task, 'duration_secs': 0.143428} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.593737] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287758', 'volume_id': '2540a76c-67a2-4e5f-853c-f70f7cef4420', 'name': 'volume-2540a76c-67a2-4e5f-853c-f70f7cef4420', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50', 'attached_at': '', 'detached_at': '', 'volume_id': '2540a76c-67a2-4e5f-853c-f70f7cef4420', 'serial': '2540a76c-67a2-4e5f-853c-f70f7cef4420'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1130.632292] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52fd2792-6a8d-4510-9035-4e67a7ce8dae, 'name': SearchDatastore_Task, 'duration_secs': 0.009355} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.632587] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.632814] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1130.633064] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.633220] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.633400] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1130.633661] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c86c6758-5e20-4eba-9b09-7b4d6aa01c6a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.641393] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1130.641573] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1130.642258] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e782617-4dbd-4e64-bef9-b89530eddc1e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.647253] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1130.647253] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52be2465-b521-6c9d-5a93-3b809e5143b9" [ 1130.647253] env[63355]: _type = "Task" [ 1130.647253] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.654273] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52be2465-b521-6c9d-5a93-3b809e5143b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.882629] env[63355]: INFO nova.compute.manager [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Took 11.99 seconds to build instance. [ 1130.949989] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350196, 'name': ReconfigVM_Task, 'duration_secs': 0.348975} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.950234] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe/4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1130.950878] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e723fb7c-80e5-4684-9f4b-431831a4f79e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.957336] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for the task: (returnval){ [ 1130.957336] env[63355]: value = "task-1350197" [ 1130.957336] env[63355]: _type = "Task" [ 1130.957336] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.966222] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350197, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.973804] env[63355]: DEBUG oslo_concurrency.lockutils [req-1b01a39b-c0f7-4712-9152-539f56a12b93 req-115a3494-67f6-4ba9-bb7a-5c4cc2fa3793 service nova] Releasing lock "refresh_cache-8d332e8f-4c89-4050-b5b7-3cd034f9edfe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.158194] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52be2465-b521-6c9d-5a93-3b809e5143b9, 'name': SearchDatastore_Task, 'duration_secs': 0.008842} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.159128] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f13cee4-971b-4735-b008-48e8501f4101 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.164525] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1131.164525] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5240d6bd-a660-878a-0092-fd652c4c2a93" [ 1131.164525] env[63355]: _type = "Task" [ 1131.164525] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.172246] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5240d6bd-a660-878a-0092-fd652c4c2a93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.385035] env[63355]: DEBUG oslo_concurrency.lockutils [None req-5eeec667-e6cc-4856-a473-983b09dfb823 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "8db23054-4556-496a-a4b1-a24c71c30f50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.505s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.467788] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350197, 'name': Rename_Task, 'duration_secs': 0.153943} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.468154] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1131.468417] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d2cd59e-0dc2-4b72-9b42-6508077bc160 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.475033] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for the task: (returnval){ [ 1131.475033] env[63355]: value = "task-1350198" [ 1131.475033] env[63355]: _type = "Task" [ 1131.475033] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.483064] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350198, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.634074] env[63355]: DEBUG nova.objects.instance [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'flavor' on Instance uuid 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.675036] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5240d6bd-a660-878a-0092-fd652c4c2a93, 'name': SearchDatastore_Task, 'duration_secs': 0.00975} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.675273] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.675533] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 8d332e8f-4c89-4050-b5b7-3cd034f9edfe/8d332e8f-4c89-4050-b5b7-3cd034f9edfe.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1131.675819] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d7ee67e1-5a86-4294-b292-acea7496c0f4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.682521] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1131.682521] env[63355]: value = "task-1350199" [ 1131.682521] env[63355]: _type = "Task" [ 1131.682521] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.689991] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350199, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.989252] env[63355]: DEBUG oslo_vmware.api [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350198, 'name': PowerOnVM_Task, 'duration_secs': 0.484417} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.990089] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1131.990412] env[63355]: INFO nova.compute.manager [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Took 6.72 seconds to spawn the instance on the hypervisor. [ 1131.990659] env[63355]: DEBUG nova.compute.manager [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1131.991626] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad601264-9ba7-45d2-bee5-e06b7b6ae8d6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.130272] env[63355]: DEBUG nova.compute.manager [req-aec87d24-3a47-44e4-b714-820e3b9b3c60 req-c58ef286-01de-495e-b98f-284a7e202b40 service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Received event network-changed-e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1132.130909] env[63355]: DEBUG nova.compute.manager [req-aec87d24-3a47-44e4-b714-820e3b9b3c60 req-c58ef286-01de-495e-b98f-284a7e202b40 service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Refreshing instance network info cache due to event network-changed-e3455ba6-ed91-4901-b232-76ca673421a7. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1132.131193] env[63355]: DEBUG oslo_concurrency.lockutils [req-aec87d24-3a47-44e4-b714-820e3b9b3c60 req-c58ef286-01de-495e-b98f-284a7e202b40 service nova] Acquiring lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.131369] env[63355]: DEBUG oslo_concurrency.lockutils [req-aec87d24-3a47-44e4-b714-820e3b9b3c60 req-c58ef286-01de-495e-b98f-284a7e202b40 service nova] Acquired lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.131560] env[63355]: DEBUG nova.network.neutron [req-aec87d24-3a47-44e4-b714-820e3b9b3c60 req-c58ef286-01de-495e-b98f-284a7e202b40 service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Refreshing network info cache for port e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1132.139182] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4a42f000-7364-49e7-9a59-2868f4e75290 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.758s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.192050] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350199, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476123} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.192346] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 8d332e8f-4c89-4050-b5b7-3cd034f9edfe/8d332e8f-4c89-4050-b5b7-3cd034f9edfe.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1132.192562] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1132.192812] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f8f1447-c098-4ed9-8932-915be0aacc18 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.199260] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1132.199260] env[63355]: value = "task-1350200" [ 1132.199260] env[63355]: _type = "Task" [ 1132.199260] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.207072] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350200, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.480803] env[63355]: DEBUG nova.compute.manager [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Stashing vm_state: active {{(pid=63355) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1132.514517] env[63355]: INFO nova.compute.manager [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Took 12.33 seconds to build instance. [ 1132.712158] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350200, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05267} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.712515] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1132.713320] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e28bcf-a272-4b58-944e-411b972f4459 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.735099] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 8d332e8f-4c89-4050-b5b7-3cd034f9edfe/8d332e8f-4c89-4050-b5b7-3cd034f9edfe.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1132.737553] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16792132-4fcd-47f8-8553-bfe06631ea84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.758461] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1132.758461] env[63355]: value = "task-1350201" [ 1132.758461] env[63355]: _type = "Task" [ 1132.758461] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.766668] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350201, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.890262] env[63355]: DEBUG nova.network.neutron [req-aec87d24-3a47-44e4-b714-820e3b9b3c60 req-c58ef286-01de-495e-b98f-284a7e202b40 service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updated VIF entry in instance network info cache for port e3455ba6-ed91-4901-b232-76ca673421a7. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1132.890648] env[63355]: DEBUG nova.network.neutron [req-aec87d24-3a47-44e4-b714-820e3b9b3c60 req-c58ef286-01de-495e-b98f-284a7e202b40 service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updating instance_info_cache with network_info: [{"id": "e3455ba6-ed91-4901-b232-76ca673421a7", "address": "fa:16:3e:eb:3b:b7", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3455ba6-ed", "ovs_interfaceid": "e3455ba6-ed91-4901-b232-76ca673421a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.001490] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.001932] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.017497] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ee6ef078-13c5-4d20-9ffa-18cbd5295adb tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.836s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.133792] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1133.134054] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1133.134248] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1133.268562] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350201, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.393512] env[63355]: DEBUG oslo_concurrency.lockutils [req-aec87d24-3a47-44e4-b714-820e3b9b3c60 req-c58ef286-01de-495e-b98f-284a7e202b40 service nova] Releasing lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.507093] env[63355]: INFO nova.compute.claims [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1133.563943] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquiring lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.564445] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.564537] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquiring lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.564770] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.565016] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.568532] env[63355]: INFO nova.compute.manager [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Terminating instance [ 1133.570403] env[63355]: DEBUG nova.compute.manager [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1133.570606] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1133.571446] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430ba2a2-901c-44fc-a29b-1f0f670f9628 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.578757] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1133.578981] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e654b121-1a49-41d9-bf41-10eb9f01066d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.584881] env[63355]: DEBUG oslo_vmware.api [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for the task: (returnval){ [ 1133.584881] env[63355]: value = "task-1350202" [ 1133.584881] env[63355]: _type = "Task" [ 1133.584881] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.592171] env[63355]: DEBUG oslo_vmware.api [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350202, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.769454] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350201, 'name': ReconfigVM_Task, 'duration_secs': 0.931978} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.769713] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 8d332e8f-4c89-4050-b5b7-3cd034f9edfe/8d332e8f-4c89-4050-b5b7-3cd034f9edfe.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1133.770337] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be42295f-4e07-4a00-827d-fe25a2c75ca7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.776733] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1133.776733] env[63355]: value = "task-1350203" [ 1133.776733] env[63355]: _type = "Task" [ 1133.776733] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.784605] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350203, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.013753] env[63355]: INFO nova.compute.resource_tracker [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating resource usage from migration a61bc71f-037c-404e-a40f-bdce48d14940 [ 1134.096209] env[63355]: DEBUG oslo_vmware.api [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350202, 'name': PowerOffVM_Task, 'duration_secs': 0.278436} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.096494] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1134.096667] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1134.096920] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f752e3e3-a5fe-40a9-b6f0-545661902186 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.099866] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c8c9f7-e21e-46d4-b0b2-3d186f009736 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.106528] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89cb2a7-d9e3-4965-a64d-9cc74e578567 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.137659] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47aceb50-f980-4138-957a-20c291160f52 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.141285] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.141420] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquired lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.141563] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Forcefully refreshing network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1134.147217] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119adaad-b950-4362-9edc-9b1fc791eb88 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.161331] env[63355]: DEBUG nova.compute.provider_tree [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.163360] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1134.163570] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1134.163755] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Deleting the datastore file [datastore2] 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1134.164197] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2c2479c-2d57-430a-b1ce-34144be3b43e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.170902] env[63355]: DEBUG oslo_vmware.api [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for the task: (returnval){ [ 1134.170902] env[63355]: value = "task-1350205" [ 1134.170902] env[63355]: _type = "Task" [ 1134.170902] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.178728] env[63355]: DEBUG oslo_vmware.api [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350205, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.286747] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350203, 'name': Rename_Task, 'duration_secs': 0.290256} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.286994] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1134.287274] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-adafcaf4-59f1-4b0b-947b-767741d25805 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.294266] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1134.294266] env[63355]: value = "task-1350206" [ 1134.294266] env[63355]: _type = "Task" [ 1134.294266] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.301925] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.666237] env[63355]: DEBUG nova.scheduler.client.report [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1134.682643] env[63355]: DEBUG oslo_vmware.api [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Task: {'id': task-1350205, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186976} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.683512] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1134.683713] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1134.683893] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1134.684090] env[63355]: INFO nova.compute.manager [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1134.684346] env[63355]: DEBUG oslo.service.loopingcall [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1134.684962] env[63355]: DEBUG nova.compute.manager [-] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1134.685080] env[63355]: DEBUG nova.network.neutron [-] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1134.804828] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350206, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.963864] env[63355]: DEBUG nova.compute.manager [req-8b895b1b-7d5f-484d-bfb8-c1f71fc3960c req-843e07cc-73bb-43c8-bf01-fa11371ffec9 service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Received event network-vif-deleted-8c03f4b6-4e4b-4300-932c-1c26be9d108f {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1134.963864] env[63355]: INFO nova.compute.manager [req-8b895b1b-7d5f-484d-bfb8-c1f71fc3960c req-843e07cc-73bb-43c8-bf01-fa11371ffec9 service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Neutron deleted interface 8c03f4b6-4e4b-4300-932c-1c26be9d108f; detaching it from the instance and deleting it from the info cache [ 1134.963864] env[63355]: DEBUG nova.network.neutron [req-8b895b1b-7d5f-484d-bfb8-c1f71fc3960c req-843e07cc-73bb-43c8-bf01-fa11371ffec9 service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.170841] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.169s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.171074] env[63355]: INFO nova.compute.manager [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Migrating [ 1135.304094] env[63355]: DEBUG oslo_vmware.api [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350206, 'name': PowerOnVM_Task, 'duration_secs': 0.512182} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.304400] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1135.304605] env[63355]: INFO nova.compute.manager [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Took 7.87 seconds to spawn the instance on the hypervisor. [ 1135.304797] env[63355]: DEBUG nova.compute.manager [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1135.307688] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538fd63f-0db4-490b-8916-9fa4420cbe65 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.374825] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance_info_cache with network_info: [{"id": "2689b07f-15d6-4cad-b652-24c86298c985", "address": "fa:16:3e:cf:90:6f", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2689b07f-15", "ovs_interfaceid": "2689b07f-15d6-4cad-b652-24c86298c985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.435105] env[63355]: DEBUG nova.network.neutron [-] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.466642] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-222b2cb6-88fd-4cd1-aafe-7ac17c6ba4dc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.474786] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f17af59-79a9-4384-8e2c-e6c8af989069 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.499999] env[63355]: DEBUG nova.compute.manager [req-8b895b1b-7d5f-484d-bfb8-c1f71fc3960c req-843e07cc-73bb-43c8-bf01-fa11371ffec9 service nova] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Detach interface failed, port_id=8c03f4b6-4e4b-4300-932c-1c26be9d108f, reason: Instance 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1135.687914] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.825979] env[63355]: INFO nova.compute.manager [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Took 12.99 seconds to build instance. [ 1135.877643] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Releasing lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.877643] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updated the network info_cache for instance {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1135.877869] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.878090] env[63355]: DEBUG nova.network.neutron [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1135.879164] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.879533] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.879698] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.879853] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.879999] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.880168] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.880302] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1135.880440] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.941088] env[63355]: INFO nova.compute.manager [-] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Took 1.26 seconds to deallocate network for instance. [ 1135.991946] env[63355]: DEBUG nova.compute.manager [req-ee423d8d-7fe3-46a8-8b34-99991e9f0e62 req-d54ac295-9e6b-4589-94d4-f43ba458f225 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Received event network-changed-69950442-ccce-4a0a-95ad-cc51720ad0a6 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1135.991946] env[63355]: DEBUG nova.compute.manager [req-ee423d8d-7fe3-46a8-8b34-99991e9f0e62 req-d54ac295-9e6b-4589-94d4-f43ba458f225 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Refreshing instance network info cache due to event network-changed-69950442-ccce-4a0a-95ad-cc51720ad0a6. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1135.992179] env[63355]: DEBUG oslo_concurrency.lockutils [req-ee423d8d-7fe3-46a8-8b34-99991e9f0e62 req-d54ac295-9e6b-4589-94d4-f43ba458f225 service nova] Acquiring lock "refresh_cache-8d332e8f-4c89-4050-b5b7-3cd034f9edfe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.992328] env[63355]: DEBUG oslo_concurrency.lockutils [req-ee423d8d-7fe3-46a8-8b34-99991e9f0e62 req-d54ac295-9e6b-4589-94d4-f43ba458f225 service nova] Acquired lock "refresh_cache-8d332e8f-4c89-4050-b5b7-3cd034f9edfe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.992495] env[63355]: DEBUG nova.network.neutron [req-ee423d8d-7fe3-46a8-8b34-99991e9f0e62 req-d54ac295-9e6b-4589-94d4-f43ba458f225 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Refreshing network info cache for port 69950442-ccce-4a0a-95ad-cc51720ad0a6 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1136.327822] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a014ddaa-c7d3-4982-a4aa-861302276e90 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.498s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.386043] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.386043] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.386043] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.386043] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1136.387445] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3880dc8f-94fa-41cc-93ea-15f64cde29b4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.397514] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc3dc8a-3cf7-4c01-a7fc-ea3362dd3e34 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.420290] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd45b19c-2589-4486-a66e-ffc322928eaa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.428093] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1effe0-c20c-4c38-98c2-6feb2f857905 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.456753] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.457013] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.457244] env[63355]: DEBUG nova.objects.instance [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lazy-loading 'resources' on Instance uuid 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.458314] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180886MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1136.458451] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.692823] env[63355]: DEBUG nova.network.neutron [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance_info_cache with network_info: [{"id": "2689b07f-15d6-4cad-b652-24c86298c985", "address": "fa:16:3e:cf:90:6f", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2689b07f-15", "ovs_interfaceid": "2689b07f-15d6-4cad-b652-24c86298c985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.755467] env[63355]: DEBUG nova.network.neutron [req-ee423d8d-7fe3-46a8-8b34-99991e9f0e62 req-d54ac295-9e6b-4589-94d4-f43ba458f225 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Updated VIF entry in instance network info cache for port 69950442-ccce-4a0a-95ad-cc51720ad0a6. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1136.755840] env[63355]: DEBUG nova.network.neutron [req-ee423d8d-7fe3-46a8-8b34-99991e9f0e62 req-d54ac295-9e6b-4589-94d4-f43ba458f225 service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Updating instance_info_cache with network_info: [{"id": "69950442-ccce-4a0a-95ad-cc51720ad0a6", "address": "fa:16:3e:97:4f:85", "network": {"id": "fa8fe114-ece5-4e31-9b0d-b3a4e7fe2bbe", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1167142504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58d215a4ff55488f931814352915d256", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69950442-cc", "ovs_interfaceid": "69950442-ccce-4a0a-95ad-cc51720ad0a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.122591] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0188d5-e5db-405d-a8ed-f4214a5ce82f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.130383] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebd2c65-b65c-459f-b534-89cffcdb461d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.160533] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f182d33d-5500-4888-a6f4-67eb219405d1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.167627] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54036423-f517-4b96-8be2-723822913653 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.180472] env[63355]: DEBUG nova.compute.provider_tree [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.195190] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.258055] env[63355]: DEBUG oslo_concurrency.lockutils [req-ee423d8d-7fe3-46a8-8b34-99991e9f0e62 req-d54ac295-9e6b-4589-94d4-f43ba458f225 service nova] Releasing lock "refresh_cache-8d332e8f-4c89-4050-b5b7-3cd034f9edfe" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.684185] env[63355]: DEBUG nova.scheduler.client.report [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1138.188886] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.732s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.191189] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.733s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.204782] env[63355]: INFO nova.scheduler.client.report [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Deleted allocations for instance 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe [ 1138.707910] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a3df39-717f-4f53-9687-42d3057b17b3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.714233] env[63355]: DEBUG oslo_concurrency.lockutils [None req-97a08c6e-260f-4fb5-a502-6bb7803debf9 tempest-ServersNegativeTestMultiTenantJSON-245664265 tempest-ServersNegativeTestMultiTenantJSON-245664265-project-member] Lock "4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.150s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.732296] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50' progress to 0 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1139.199906] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Applying migration context for instance 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50 as it has an incoming, in-progress migration a61bc71f-037c-404e-a40f-bdce48d14940. Migration status is migrating {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1139.200833] env[63355]: INFO nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating resource usage from migration a61bc71f-037c-404e-a40f-bdce48d14940 [ 1139.220387] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1139.220534] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8db23054-4556-496a-a4b1-a24c71c30f50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1139.220659] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8d332e8f-4c89-4050-b5b7-3cd034f9edfe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1139.220788] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Migration a61bc71f-037c-404e-a40f-bdce48d14940 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1139.220906] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1139.221129] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1139.221298] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1139.237944] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1139.241448] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6624717d-6357-4e5d-a62c-f5ae61ce3a2d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.248262] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1139.248262] env[63355]: value = "task-1350207" [ 1139.248262] env[63355]: _type = "Task" [ 1139.248262] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.260773] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350207, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.305150] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39fb262d-06bb-4b2f-b150-ab2d6165198d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.313611] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a1c34b-9c18-4c9d-8d14-5efe8a5d74bf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.347564] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6de7b1-57d0-44f7-802c-e9ae93e1dc15 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.356086] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cdf8ad-f9d5-4667-9a44-ccf611bcaf2f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.369705] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.759265] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350207, 'name': PowerOffVM_Task, 'duration_secs': 0.192204} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.759545] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1139.760308] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50' progress to 17 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1139.873525] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1140.266630] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1140.266937] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1140.267036] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1140.267234] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1140.267387] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1140.267536] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1140.267741] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1140.267902] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1140.268238] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1140.268458] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1140.268647] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1140.277141] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85def5c5-aabb-4242-8663-379778ccf4f4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.294331] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1140.294331] env[63355]: value = "task-1350208" [ 1140.294331] env[63355]: _type = "Task" [ 1140.294331] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.302398] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350208, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.377766] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1140.377988] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.187s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.378264] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1140.378406] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Cleaning up deleted instances {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1140.805691] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350208, 'name': ReconfigVM_Task, 'duration_secs': 0.461169} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.805899] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50' progress to 33 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1140.890727] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] There are 43 instances to clean {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1140.891016] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 4ae98a75-a5a9-4f00-9fdf-51a6eff06fbe] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1141.312419] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1141.312895] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1141.313105] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1141.313333] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1141.313463] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1141.313619] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1141.313852] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1141.313983] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1141.315025] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1141.315025] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1141.315025] env[63355]: DEBUG nova.virt.hardware [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1141.319725] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1141.320023] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2c2dd65-363d-4dbe-9841-620bc24222e0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.338887] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1141.338887] env[63355]: value = "task-1350209" [ 1141.338887] env[63355]: _type = "Task" [ 1141.338887] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.347373] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350209, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.394054] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: bcf935ad-afa2-4b36-b01b-14020453ad98] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1141.849077] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350209, 'name': ReconfigVM_Task, 'duration_secs': 0.184368} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.849420] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1141.850219] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca76972-bd57-43ce-9bfb-55998c4c7312 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.874538] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50/8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1141.874836] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a901b89b-141e-4cb6-8801-492bf2fb4734 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.891924] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1141.891924] env[63355]: value = "task-1350210" [ 1141.891924] env[63355]: _type = "Task" [ 1141.891924] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.899567] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8fc3ea4e-d525-4583-b237-f895545fd3d1] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1141.901398] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350210, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.401504] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350210, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.403079] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: f08bca27-ab50-40d0-a6d9-4842a37a4faf] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1142.901940] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350210, 'name': ReconfigVM_Task, 'duration_secs': 0.867408} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.903028] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50/8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1142.903028] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50' progress to 50 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1142.906171] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 50e7499c-a8c1-41a6-adac-36e6c3b92819] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1143.410051] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: e398ee67-eb02-4256-9120-06111f110692] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1143.412527] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77ffd39-4d7c-4481-be8d-759c93004963 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.436060] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed7d756-833a-42b5-84b8-e9fca4fdf385 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.456090] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50' progress to 67 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1143.916204] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: e271159e-14ed-4a88-a7b8-23d404d77985] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1144.419437] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: fa748748-31ec-431a-a628-5ea179e26fc8] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1144.922409] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 2214ffc8-d0fd-49f3-91c1-74c13ef7bc07] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1145.091409] env[63355]: DEBUG nova.network.neutron [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Port 2689b07f-15d6-4cad-b652-24c86298c985 binding to destination host cpu-1 is already ACTIVE {{(pid=63355) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1145.425908] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 40c48196-6543-4c96-9b8c-3b3c9fbccfe4] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1145.929131] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: cf2d004b-b9dd-4c05-b54d-9509cd1d0b06] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1146.109213] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.109452] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.109635] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.431946] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 9bf49da7-df44-4f26-ac7d-d3a4dab24ce7] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1146.936059] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 9ee3001b-8bf0-43ab-996a-a68dad57d8e6] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1147.149419] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.149661] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.149962] env[63355]: DEBUG nova.network.neutron [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1147.439308] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: da3cb83c-6368-49c0-9b11-0498221e3c0f] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1147.875711] env[63355]: DEBUG nova.network.neutron [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance_info_cache with network_info: [{"id": "2689b07f-15d6-4cad-b652-24c86298c985", "address": "fa:16:3e:cf:90:6f", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2689b07f-15", "ovs_interfaceid": "2689b07f-15d6-4cad-b652-24c86298c985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.942591] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: edb14fe7-d444-4fef-8c5d-d5616676e326] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1148.378522] env[63355]: DEBUG oslo_concurrency.lockutils [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.445648] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: fd55ddfc-f376-4e15-961e-6826f31a7890] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1148.888204] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31b23de-b125-43c4-8211-56d2b0df25f4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.895342] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85eb7b39-b963-4c8b-a82c-3349639499d2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.949205] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: c630c5fe-6907-4952-9807-6e59bd1cc9e1] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1149.070366] env[63355]: DEBUG oslo_concurrency.lockutils [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.070649] env[63355]: DEBUG oslo_concurrency.lockutils [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.453597] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: ae3961d2-dc5b-4e49-acca-6fb52291f23b] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1149.574027] env[63355]: DEBUG nova.compute.utils [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1149.956341] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 248ec2f2-3a41-444a-b550-71c5438a45f7] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1149.998400] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a61228e-81a6-4eaf-8551-f0b39cffd251 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.019569] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d8297d-e03c-4c12-9267-820199d3011c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.026398] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50' progress to 83 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1150.076379] env[63355]: DEBUG oslo_concurrency.lockutils [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.459637] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8e9ff9b7-636f-48df-9168-509d733278f8] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1150.532512] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1150.532759] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa627568-124c-4d15-af10-2909ac4fa922 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.540779] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1150.540779] env[63355]: value = "task-1350211" [ 1150.540779] env[63355]: _type = "Task" [ 1150.540779] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.548862] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350211, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.962484] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: fc6a99e6-5319-47d3-8175-770c06d3e325] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1151.051707] env[63355]: DEBUG oslo_vmware.api [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350211, 'name': PowerOnVM_Task, 'duration_secs': 0.368258} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.051974] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1151.052183] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-581ef886-b96c-4a29-9c75-633e60a4aa72 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50' progress to 100 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1151.141681] env[63355]: DEBUG oslo_concurrency.lockutils [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.141864] env[63355]: DEBUG oslo_concurrency.lockutils [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.141972] env[63355]: INFO nova.compute.manager [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Attaching volume 359e6f56-7a86-4095-95e0-d874eaf30d00 to /dev/sdb [ 1151.171459] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb792a6-e4ca-4a12-8e99-00645a4d42cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.178380] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089dc4e0-0159-482a-8b43-2b03e89680e5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.191042] env[63355]: DEBUG nova.virt.block_device [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Updating existing volume attachment record: ce317fb2-4cfa-4a61-97c3-b0fc4b6d5518 {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1151.465739] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 1570cb36-76e0-4d06-8080-735b5246e92e] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1151.969660] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 6706c65c-3959-471b-82c0-2196f4aa1ab0] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1152.473411] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 7c153109-b814-4e11-b4f9-7b8cebb853d4] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1152.976327] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: a408e19a-0960-430d-8550-0a304c63da61] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1153.288020] env[63355]: DEBUG nova.network.neutron [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Port 2689b07f-15d6-4cad-b652-24c86298c985 binding to destination host cpu-1 is already ACTIVE {{(pid=63355) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1153.288402] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.288611] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.288795] env[63355]: DEBUG nova.network.neutron [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1153.479295] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: d2480bd6-527b-46b8-8a6a-22ad4eda3f5f] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1153.982881] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: ecbacdb3-516e-43c3-96e3-4961b76565ca] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1153.992952] env[63355]: DEBUG nova.network.neutron [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance_info_cache with network_info: [{"id": "2689b07f-15d6-4cad-b652-24c86298c985", "address": "fa:16:3e:cf:90:6f", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2689b07f-15", "ovs_interfaceid": "2689b07f-15d6-4cad-b652-24c86298c985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.487110] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 199bc488-2e5d-4cea-aefb-ddd35ecc7a30] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1154.495357] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.990203] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: dc56b350-cee7-49c4-9712-8c8c29146ff2] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1154.998840] env[63355]: DEBUG nova.compute.manager [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63355) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1155.493933] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 537d1cfc-7c1d-4a3b-909d-ed6b20c42ae5] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1155.733516] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1155.733781] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287767', 'volume_id': '359e6f56-7a86-4095-95e0-d874eaf30d00', 'name': 'volume-359e6f56-7a86-4095-95e0-d874eaf30d00', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7aee05be-b8b4-4767-b5d4-88aa4a21d5cd', 'attached_at': '', 'detached_at': '', 'volume_id': '359e6f56-7a86-4095-95e0-d874eaf30d00', 'serial': '359e6f56-7a86-4095-95e0-d874eaf30d00'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1155.734719] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e24f36a-8d0b-432a-9aac-b8d3388f962e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.750842] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f525eb8d-3c9e-43a6-a6a3-28bb8920b2a4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.774360] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] volume-359e6f56-7a86-4095-95e0-d874eaf30d00/volume-359e6f56-7a86-4095-95e0-d874eaf30d00.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1155.774645] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc3eced5-59aa-449c-a3c3-535b7396bf3c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.792182] env[63355]: DEBUG oslo_vmware.api [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1155.792182] env[63355]: value = "task-1350216" [ 1155.792182] env[63355]: _type = "Task" [ 1155.792182] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.799815] env[63355]: DEBUG oslo_vmware.api [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350216, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.997355] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 74071e73-10fa-4dcb-aa15-91303b2278fb] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1156.090989] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.091427] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.303931] env[63355]: DEBUG oslo_vmware.api [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350216, 'name': ReconfigVM_Task, 'duration_secs': 0.324922} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.304176] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Reconfigured VM instance instance-00000065 to attach disk [datastore1] volume-359e6f56-7a86-4095-95e0-d874eaf30d00/volume-359e6f56-7a86-4095-95e0-d874eaf30d00.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1156.308656] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b90c091-e1f7-4b03-a496-50a51175447b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.323747] env[63355]: DEBUG oslo_vmware.api [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1156.323747] env[63355]: value = "task-1350217" [ 1156.323747] env[63355]: _type = "Task" [ 1156.323747] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.330891] env[63355]: DEBUG oslo_vmware.api [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350217, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.500445] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 7c139710-d8d9-4cd7-bec0-6e021d3b2e68] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1156.594427] env[63355]: DEBUG nova.objects.instance [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'migration_context' on Instance uuid 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1156.834239] env[63355]: DEBUG oslo_vmware.api [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350217, 'name': ReconfigVM_Task, 'duration_secs': 0.133098} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.834239] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287767', 'volume_id': '359e6f56-7a86-4095-95e0-d874eaf30d00', 'name': 'volume-359e6f56-7a86-4095-95e0-d874eaf30d00', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7aee05be-b8b4-4767-b5d4-88aa4a21d5cd', 'attached_at': '', 'detached_at': '', 'volume_id': '359e6f56-7a86-4095-95e0-d874eaf30d00', 'serial': '359e6f56-7a86-4095-95e0-d874eaf30d00'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1157.004221] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 2539a79e-01c0-4e0c-aa66-8784441c6fda] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1157.170309] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d37d01d5-a400-4c31-95c2-33b44e935d6d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.177895] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5491b14-cb61-462c-9c32-c846731e02ff {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.206972] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352a1069-6d47-4992-8689-5d54b96ffe4d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.213777] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa722ae-c0a5-44ca-894c-ad6c7a87dc38 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.227249] env[63355]: DEBUG nova.compute.provider_tree [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1157.506852] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: b4b09b1d-680e-47b8-aa8a-9b3d9167824d] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1157.730079] env[63355]: DEBUG nova.scheduler.client.report [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1157.867431] env[63355]: DEBUG nova.objects.instance [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lazy-loading 'flavor' on Instance uuid 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.009825] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: b6056441-9ee4-484f-a1d2-077546f2c581] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1158.375222] env[63355]: DEBUG oslo_concurrency.lockutils [None req-03caf074-404b-471d-8299-a98a3ba1a64e tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.233s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.512754] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8e3ec9d3-bc22-4e39-ad7c-93268dd59020] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1158.740691] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.649s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.016313] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: f5e62ce1-40b7-4648-a4a6-068ff06eaf9b] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1159.496593] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "ddaab84c-5f1d-4122-9408-2043f30a2227" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.496823] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "ddaab84c-5f1d-4122-9408-2043f30a2227" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.519498] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: b8e153d5-a2dc-43e6-beb8-59a7cc0ddbda] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1159.999017] env[63355]: DEBUG nova.compute.manager [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1160.023128] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 0b660647-697a-445b-bca4-82ef6975ea75] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1160.278627] env[63355]: INFO nova.compute.manager [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Swapping old allocation on dict_keys(['47abb610-db7e-4770-911d-187dd075ef8b']) held by migration a61bc71f-037c-404e-a40f-bdce48d14940 for instance [ 1160.298998] env[63355]: DEBUG nova.scheduler.client.report [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Overwriting current allocation {'allocations': {'47abb610-db7e-4770-911d-187dd075ef8b': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 138}}, 'project_id': '6b3fee9bc99d49ea9de53d5dce52c79d', 'user_id': '54ad2682257b4250a8a96f4ba6a9523b', 'consumer_generation': 1} on consumer 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50 {{(pid=63355) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1160.390342] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.390663] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.390938] env[63355]: DEBUG nova.network.neutron [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1160.517143] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.517380] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.518948] env[63355]: INFO nova.compute.claims [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1160.527202] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 386e847e-967b-4247-9730-cdc5ac251474] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1161.030187] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 5591bb55-83d7-4301-a3f9-fde945632344] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1161.107061] env[63355]: DEBUG nova.network.neutron [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance_info_cache with network_info: [{"id": "2689b07f-15d6-4cad-b652-24c86298c985", "address": "fa:16:3e:cf:90:6f", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2689b07f-15", "ovs_interfaceid": "2689b07f-15d6-4cad-b652-24c86298c985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.534011] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 3d918cab-6ed6-4a37-a024-28e3db1b779c] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1161.607790] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5876f1a-d894-465a-8bfc-f5c271574138 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.610722] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.611535] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add0fb55-dac3-4b8a-8e3f-259974836dce {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.619936] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af2ba59-d6cc-4996-aca0-268433fbecca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.624299] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c993f1-3913-4319-9104-7349dc0bb1f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.658091] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ad699d-5514-4137-b4f1-67f8d28a6036 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.667145] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9364c4f9-b05a-4d9d-a654-e7c0fb96a017 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.679805] env[63355]: DEBUG nova.compute.provider_tree [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.037089] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 9989c129-07d3-4af9-9a86-1e2746d6ac70] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1162.183203] env[63355]: DEBUG nova.scheduler.client.report [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1162.541054] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1162.541054] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Cleaning up deleted instances with incomplete migration {{(pid=63355) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1162.688170] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.688813] env[63355]: DEBUG nova.compute.manager [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1162.727791] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1162.728112] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e9515ad-f775-4923-b4fe-104fba3beb8f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.735571] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1162.735571] env[63355]: value = "task-1350218" [ 1162.735571] env[63355]: _type = "Task" [ 1162.735571] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.743551] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350218, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.043242] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.193625] env[63355]: DEBUG nova.compute.utils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1163.194952] env[63355]: DEBUG nova.compute.manager [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1163.195135] env[63355]: DEBUG nova.network.neutron [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1163.241625] env[63355]: DEBUG nova.policy [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7302e68a14984b08aef9c3af0d2a12ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '876995f179ed46b397822fa1be08ea29', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1163.246167] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350218, 'name': PowerOffVM_Task, 'duration_secs': 0.254209} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.246526] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1163.247191] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1163.247410] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1163.247572] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1163.247754] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1163.247903] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1163.248066] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1163.248274] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1163.248460] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1163.248635] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1163.248803] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1163.248975] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1163.253814] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-783ff352-3da3-401c-ae86-e5081492b2a3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.268802] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1163.268802] env[63355]: value = "task-1350219" [ 1163.268802] env[63355]: _type = "Task" [ 1163.268802] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.277842] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350219, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.489209] env[63355]: DEBUG nova.network.neutron [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Successfully created port: bba4c0ca-7f31-49ce-92ec-703f913914ce {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1163.697920] env[63355]: DEBUG nova.compute.manager [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1163.777744] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350219, 'name': ReconfigVM_Task, 'duration_secs': 0.140265} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.778629] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa81e33-f104-43be-90dd-0783c594e9d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.807778] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1163.808067] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1163.808293] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1163.808660] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1163.808832] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1163.809066] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1163.809365] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1163.809706] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1163.809938] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1163.810139] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1163.810333] env[63355]: DEBUG nova.virt.hardware [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1163.811150] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b0c1595-2f11-41da-8724-ca567a0cc4aa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.816753] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1163.816753] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52dd65bb-ee3c-9def-7845-593209e1ec72" [ 1163.816753] env[63355]: _type = "Task" [ 1163.816753] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.824602] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52dd65bb-ee3c-9def-7845-593209e1ec72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.327480] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52dd65bb-ee3c-9def-7845-593209e1ec72, 'name': SearchDatastore_Task, 'duration_secs': 0.014427} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.333038] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1164.333318] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9bcdb5f1-4e95-4afc-b611-2b1b68f2c0d7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.351509] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1164.351509] env[63355]: value = "task-1350220" [ 1164.351509] env[63355]: _type = "Task" [ 1164.351509] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.358759] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350220, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.707342] env[63355]: DEBUG nova.compute.manager [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1164.732599] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1164.732851] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1164.733021] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1164.733213] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1164.733366] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1164.733517] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1164.733727] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1164.734301] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1164.734301] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1164.734301] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1164.734493] env[63355]: DEBUG nova.virt.hardware [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1164.735258] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08dcede0-1e78-462f-bcf4-daa07af2377a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.742941] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92dde788-ce96-43d6-9d56-111845e27aa1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.816903] env[63355]: DEBUG nova.compute.manager [req-5273935c-b6fe-416c-be0d-057ce6f488b3 req-54210737-8919-4b09-94b3-6ccd0ab938f6 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Received event network-vif-plugged-bba4c0ca-7f31-49ce-92ec-703f913914ce {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1164.817144] env[63355]: DEBUG oslo_concurrency.lockutils [req-5273935c-b6fe-416c-be0d-057ce6f488b3 req-54210737-8919-4b09-94b3-6ccd0ab938f6 service nova] Acquiring lock "ddaab84c-5f1d-4122-9408-2043f30a2227-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.817361] env[63355]: DEBUG oslo_concurrency.lockutils [req-5273935c-b6fe-416c-be0d-057ce6f488b3 req-54210737-8919-4b09-94b3-6ccd0ab938f6 service nova] Lock "ddaab84c-5f1d-4122-9408-2043f30a2227-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.817570] env[63355]: DEBUG oslo_concurrency.lockutils [req-5273935c-b6fe-416c-be0d-057ce6f488b3 req-54210737-8919-4b09-94b3-6ccd0ab938f6 service nova] Lock "ddaab84c-5f1d-4122-9408-2043f30a2227-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.817752] env[63355]: DEBUG nova.compute.manager [req-5273935c-b6fe-416c-be0d-057ce6f488b3 req-54210737-8919-4b09-94b3-6ccd0ab938f6 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] No waiting events found dispatching network-vif-plugged-bba4c0ca-7f31-49ce-92ec-703f913914ce {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1164.818092] env[63355]: WARNING nova.compute.manager [req-5273935c-b6fe-416c-be0d-057ce6f488b3 req-54210737-8919-4b09-94b3-6ccd0ab938f6 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Received unexpected event network-vif-plugged-bba4c0ca-7f31-49ce-92ec-703f913914ce for instance with vm_state building and task_state spawning. [ 1164.861300] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350220, 'name': ReconfigVM_Task, 'duration_secs': 0.201895} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.861571] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1164.862381] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ed4b06-2991-4b24-ac2f-711e05fc13ec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.887665] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50/8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1164.888308] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-effae184-2855-427d-8249-400ca9a27357 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.906583] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1164.906583] env[63355]: value = "task-1350221" [ 1164.906583] env[63355]: _type = "Task" [ 1164.906583] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.915111] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350221, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.370213] env[63355]: DEBUG nova.network.neutron [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Successfully updated port: bba4c0ca-7f31-49ce-92ec-703f913914ce {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1165.391988] env[63355]: DEBUG nova.compute.manager [req-bde1920f-3bf2-464a-870c-24c609fb17a8 req-1080bea4-9272-44d4-acb1-c73fecad5ceb service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Received event network-changed-bba4c0ca-7f31-49ce-92ec-703f913914ce {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1165.392223] env[63355]: DEBUG nova.compute.manager [req-bde1920f-3bf2-464a-870c-24c609fb17a8 req-1080bea4-9272-44d4-acb1-c73fecad5ceb service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Refreshing instance network info cache due to event network-changed-bba4c0ca-7f31-49ce-92ec-703f913914ce. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1165.392468] env[63355]: DEBUG oslo_concurrency.lockutils [req-bde1920f-3bf2-464a-870c-24c609fb17a8 req-1080bea4-9272-44d4-acb1-c73fecad5ceb service nova] Acquiring lock "refresh_cache-ddaab84c-5f1d-4122-9408-2043f30a2227" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.392620] env[63355]: DEBUG oslo_concurrency.lockutils [req-bde1920f-3bf2-464a-870c-24c609fb17a8 req-1080bea4-9272-44d4-acb1-c73fecad5ceb service nova] Acquired lock "refresh_cache-ddaab84c-5f1d-4122-9408-2043f30a2227" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.392783] env[63355]: DEBUG nova.network.neutron [req-bde1920f-3bf2-464a-870c-24c609fb17a8 req-1080bea4-9272-44d4-acb1-c73fecad5ceb service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Refreshing network info cache for port bba4c0ca-7f31-49ce-92ec-703f913914ce {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1165.415988] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350221, 'name': ReconfigVM_Task, 'duration_secs': 0.278689} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.416409] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50/8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1165.417381] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ad54a3-7e63-4094-9436-149da6f79514 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.437198] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4a0e49-1e74-42cc-b3bd-ebd2a6a06130 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.477507] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e587b0-8cf6-4067-9a25-8af2120a818f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.504306] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60861a8-4417-4e78-b302-34158a4f5ebd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.511815] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1165.512153] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41210faa-8873-4905-a6bd-245534bff229 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.519490] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1165.519490] env[63355]: value = "task-1350222" [ 1165.519490] env[63355]: _type = "Task" [ 1165.519490] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.527561] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350222, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.873098] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "refresh_cache-ddaab84c-5f1d-4122-9408-2043f30a2227" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.922076] env[63355]: DEBUG nova.network.neutron [req-bde1920f-3bf2-464a-870c-24c609fb17a8 req-1080bea4-9272-44d4-acb1-c73fecad5ceb service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1165.990371] env[63355]: DEBUG nova.network.neutron [req-bde1920f-3bf2-464a-870c-24c609fb17a8 req-1080bea4-9272-44d4-acb1-c73fecad5ceb service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.029095] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350222, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.493245] env[63355]: DEBUG oslo_concurrency.lockutils [req-bde1920f-3bf2-464a-870c-24c609fb17a8 req-1080bea4-9272-44d4-acb1-c73fecad5ceb service nova] Releasing lock "refresh_cache-ddaab84c-5f1d-4122-9408-2043f30a2227" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.493642] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "refresh_cache-ddaab84c-5f1d-4122-9408-2043f30a2227" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.493816] env[63355]: DEBUG nova.network.neutron [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1166.529297] env[63355]: DEBUG oslo_vmware.api [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350222, 'name': PowerOnVM_Task, 'duration_secs': 0.54476} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.529590] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1167.022879] env[63355]: DEBUG nova.network.neutron [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1167.146961] env[63355]: DEBUG nova.network.neutron [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Updating instance_info_cache with network_info: [{"id": "bba4c0ca-7f31-49ce-92ec-703f913914ce", "address": "fa:16:3e:c0:a0:5a", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba4c0ca-7f", "ovs_interfaceid": "bba4c0ca-7f31-49ce-92ec-703f913914ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.574840] env[63355]: INFO nova.compute.manager [None req-d1236055-417d-465c-97ea-45e0ba0130a0 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance to original state: 'active' [ 1167.649048] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "refresh_cache-ddaab84c-5f1d-4122-9408-2043f30a2227" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.649386] env[63355]: DEBUG nova.compute.manager [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Instance network_info: |[{"id": "bba4c0ca-7f31-49ce-92ec-703f913914ce", "address": "fa:16:3e:c0:a0:5a", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba4c0ca-7f", "ovs_interfaceid": "bba4c0ca-7f31-49ce-92ec-703f913914ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1167.649816] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:a0:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bba4c0ca-7f31-49ce-92ec-703f913914ce', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1167.657291] env[63355]: DEBUG oslo.service.loopingcall [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1167.657489] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1167.657713] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0d27643f-99c6-4fc2-b8c9-5da3334a7af5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.677275] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1167.677275] env[63355]: value = "task-1350223" [ 1167.677275] env[63355]: _type = "Task" [ 1167.677275] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.684521] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350223, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.187126] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350223, 'name': CreateVM_Task, 'duration_secs': 0.298987} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.187387] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1168.187997] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.188189] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.188538] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1168.189108] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09ff7fe9-7464-443e-b687-f8d3c1dbe51f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.193642] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1168.193642] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527f315c-5ea5-7278-c30f-e051f5c479b8" [ 1168.193642] env[63355]: _type = "Task" [ 1168.193642] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.201366] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527f315c-5ea5-7278-c30f-e051f5c479b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.705295] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527f315c-5ea5-7278-c30f-e051f5c479b8, 'name': SearchDatastore_Task, 'duration_secs': 0.009215} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.705566] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.705790] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1168.706038] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.706195] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.706385] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1168.706645] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8b55650-656a-4f00-a26d-79685120dd9a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.720677] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1168.720859] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1168.721550] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4800bb0-404b-454e-9efe-dcaa14e82e6f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.726098] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1168.726098] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526f6e08-933d-42c0-39d7-a6a07096971f" [ 1168.726098] env[63355]: _type = "Task" [ 1168.726098] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.733559] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526f6e08-933d-42c0-39d7-a6a07096971f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.030648] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.030964] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.031227] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.031450] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.032667] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.036053] env[63355]: INFO nova.compute.manager [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Terminating instance [ 1169.039030] env[63355]: DEBUG nova.compute.manager [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1169.039030] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1169.039030] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00aefbbc-2344-4bd7-839d-ed0f0e851144 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.046339] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1169.046339] env[63355]: value = "task-1350224" [ 1169.046339] env[63355]: _type = "Task" [ 1169.046339] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.054886] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350224, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.236721] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]526f6e08-933d-42c0-39d7-a6a07096971f, 'name': SearchDatastore_Task, 'duration_secs': 0.031688} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.237577] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85dfa086-915e-4571-82cd-12d4a24e1eca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.242734] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1169.242734] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52807837-c96e-af1e-2e60-b7f768beae22" [ 1169.242734] env[63355]: _type = "Task" [ 1169.242734] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.250222] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52807837-c96e-af1e-2e60-b7f768beae22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.555970] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350224, 'name': PowerOffVM_Task, 'duration_secs': 0.184791} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.556285] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1169.556516] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1169.556717] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287758', 'volume_id': '2540a76c-67a2-4e5f-853c-f70f7cef4420', 'name': 'volume-2540a76c-67a2-4e5f-853c-f70f7cef4420', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50', 'attached_at': '2024-10-24T02:34:26.000000', 'detached_at': '', 'volume_id': '2540a76c-67a2-4e5f-853c-f70f7cef4420', 'serial': '2540a76c-67a2-4e5f-853c-f70f7cef4420'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1169.557471] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70796837-9085-4f54-a1c4-3a6dc4178885 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.577746] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b26207d-a9ce-4964-bba5-724df603c7e7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.583893] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d32cebe-fea3-46d6-8081-74cbced96475 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.604320] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e0d73a-4309-4324-91f0-f4a890941c2e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.617901] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] The volume has not been displaced from its original location: [datastore2] volume-2540a76c-67a2-4e5f-853c-f70f7cef4420/volume-2540a76c-67a2-4e5f-853c-f70f7cef4420.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1169.623114] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfiguring VM instance instance-00000063 to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1169.623446] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbc45f1f-3302-4043-89bf-b3eb701703f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.641314] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1169.641314] env[63355]: value = "task-1350225" [ 1169.641314] env[63355]: _type = "Task" [ 1169.641314] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.648590] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350225, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.752973] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52807837-c96e-af1e-2e60-b7f768beae22, 'name': SearchDatastore_Task, 'duration_secs': 0.008707} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.753271] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1169.753532] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] ddaab84c-5f1d-4122-9408-2043f30a2227/ddaab84c-5f1d-4122-9408-2043f30a2227.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1169.753808] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc10cfcb-fcc1-46bc-a6dc-70da86856da1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.760397] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1169.760397] env[63355]: value = "task-1350226" [ 1169.760397] env[63355]: _type = "Task" [ 1169.760397] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.768019] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350226, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.792210] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "1c054e10-f650-4b47-a233-3bf17482a9b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.792427] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.152090] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350225, 'name': ReconfigVM_Task, 'duration_secs': 0.195156} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.152449] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Reconfigured VM instance instance-00000063 to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1170.157513] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdc00517-cac2-4e61-96e6-8711a03dc3d5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.174403] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1170.174403] env[63355]: value = "task-1350227" [ 1170.174403] env[63355]: _type = "Task" [ 1170.174403] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.183183] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350227, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.270557] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350226, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449695} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.270950] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] ddaab84c-5f1d-4122-9408-2043f30a2227/ddaab84c-5f1d-4122-9408-2043f30a2227.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1170.271059] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1170.271335] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88fcfb44-1b5e-474b-83e1-7f0b87dc8eb2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.277652] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1170.277652] env[63355]: value = "task-1350228" [ 1170.277652] env[63355]: _type = "Task" [ 1170.277652] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.284845] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350228, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.294392] env[63355]: DEBUG nova.compute.manager [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1170.683849] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350227, 'name': ReconfigVM_Task, 'duration_secs': 0.128615} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.684202] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287758', 'volume_id': '2540a76c-67a2-4e5f-853c-f70f7cef4420', 'name': 'volume-2540a76c-67a2-4e5f-853c-f70f7cef4420', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50', 'attached_at': '2024-10-24T02:34:26.000000', 'detached_at': '', 'volume_id': '2540a76c-67a2-4e5f-853c-f70f7cef4420', 'serial': '2540a76c-67a2-4e5f-853c-f70f7cef4420'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1170.684511] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1170.685261] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f60ae98-63a5-49ba-8b13-883623ad6a7d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.691674] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1170.691892] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b25fa986-0466-46fb-a260-a2630429c343 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.752573] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1170.752799] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1170.752987] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleting the datastore file [datastore1] 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.753283] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a64a05db-4ee2-4fd3-86a9-db06fd9c616f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.761185] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1170.761185] env[63355]: value = "task-1350230" [ 1170.761185] env[63355]: _type = "Task" [ 1170.761185] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.768587] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350230, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.786153] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350228, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0694} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.786409] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1170.787245] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9ff583-6e51-45cb-af8a-b24eb811a834 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.810501] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] ddaab84c-5f1d-4122-9408-2043f30a2227/ddaab84c-5f1d-4122-9408-2043f30a2227.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1170.812820] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2986cd80-70ae-41df-8483-a31a0a7b4a59 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.832670] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1170.832670] env[63355]: value = "task-1350231" [ 1170.832670] env[63355]: _type = "Task" [ 1170.832670] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.840439] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350231, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.841456] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.841684] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.843143] env[63355]: INFO nova.compute.claims [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1171.271388] env[63355]: DEBUG oslo_vmware.api [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350230, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160759} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.271795] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.271795] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1171.271947] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1171.272153] env[63355]: INFO nova.compute.manager [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1171.272401] env[63355]: DEBUG oslo.service.loopingcall [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.272595] env[63355]: DEBUG nova.compute.manager [-] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1171.272699] env[63355]: DEBUG nova.network.neutron [-] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1171.342504] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350231, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.539898] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1171.690245] env[63355]: DEBUG nova.compute.manager [req-3a0e60ec-7a0a-4cf5-b9f7-8df32e617e99 req-4f8c8907-cbff-4fd8-9aa6-26254a3dd83f service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Received event network-vif-deleted-2689b07f-15d6-4cad-b652-24c86298c985 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1171.690245] env[63355]: INFO nova.compute.manager [req-3a0e60ec-7a0a-4cf5-b9f7-8df32e617e99 req-4f8c8907-cbff-4fd8-9aa6-26254a3dd83f service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Neutron deleted interface 2689b07f-15d6-4cad-b652-24c86298c985; detaching it from the instance and deleting it from the info cache [ 1171.690354] env[63355]: DEBUG nova.network.neutron [req-3a0e60ec-7a0a-4cf5-b9f7-8df32e617e99 req-4f8c8907-cbff-4fd8-9aa6-26254a3dd83f service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.844020] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350231, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.932347] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773aaff3-4d4d-4c90-9617-96da6a911cb8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.940012] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb3969d-4d2e-40ec-bc0b-23c1a855f28e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.971893] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac637dcc-3f2f-429c-a45f-fd5fab5d8691 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.979067] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ccba7c9-0962-4062-8bd4-56433813e808 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.991575] env[63355]: DEBUG nova.compute.provider_tree [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.168396] env[63355]: DEBUG nova.network.neutron [-] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.193561] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6967226-739a-4c86-a366-2ba8b1a6422e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.203203] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f1c177-9980-4563-a2f5-ed27d63e6fee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.229096] env[63355]: DEBUG nova.compute.manager [req-3a0e60ec-7a0a-4cf5-b9f7-8df32e617e99 req-4f8c8907-cbff-4fd8-9aa6-26254a3dd83f service nova] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Detach interface failed, port_id=2689b07f-15d6-4cad-b652-24c86298c985, reason: Instance 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1172.344161] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350231, 'name': ReconfigVM_Task, 'duration_secs': 1.271006} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.344497] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Reconfigured VM instance instance-00000069 to attach disk [datastore1] ddaab84c-5f1d-4122-9408-2043f30a2227/ddaab84c-5f1d-4122-9408-2043f30a2227.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1172.345084] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba3d42e3-db0a-46b3-96af-312b303ccf73 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.351176] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1172.351176] env[63355]: value = "task-1350232" [ 1172.351176] env[63355]: _type = "Task" [ 1172.351176] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.360344] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350232, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.494231] env[63355]: DEBUG nova.scheduler.client.report [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1172.671141] env[63355]: INFO nova.compute.manager [-] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Took 1.40 seconds to deallocate network for instance. [ 1172.860855] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350232, 'name': Rename_Task, 'duration_secs': 0.137682} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.861154] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1172.861401] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ed23cb4-c62d-4ff8-a2b8-80c4559c2ed0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.867630] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1172.867630] env[63355]: value = "task-1350233" [ 1172.867630] env[63355]: _type = "Task" [ 1172.867630] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.875080] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350233, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.999282] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.157s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.999977] env[63355]: DEBUG nova.compute.manager [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1173.214045] env[63355]: INFO nova.compute.manager [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50] Took 0.54 seconds to detach 1 volumes for instance. [ 1173.378291] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350233, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.504913] env[63355]: DEBUG nova.compute.utils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1173.506345] env[63355]: DEBUG nova.compute.manager [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1173.507027] env[63355]: DEBUG nova.network.neutron [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1173.552245] env[63355]: DEBUG nova.policy [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a61bc098ea4e43ceab263ae9890da7c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff5b796e7bb0458dbd3b862dcae7bdd1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1173.692344] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.692568] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.719146] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.719399] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.719695] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.749717] env[63355]: INFO nova.scheduler.client.report [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleted allocations for instance 8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50 [ 1173.796414] env[63355]: DEBUG nova.network.neutron [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Successfully created port: 417738e8-effd-45de-9c4b-b0eda6d5ccfc {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1173.878556] env[63355]: DEBUG oslo_vmware.api [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350233, 'name': PowerOnVM_Task, 'duration_secs': 0.96227} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.878974] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1173.879254] env[63355]: INFO nova.compute.manager [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Took 9.17 seconds to spawn the instance on the hypervisor. [ 1173.879491] env[63355]: DEBUG nova.compute.manager [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1173.880285] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dcb042-df38-4f45-a1d3-24950d1a1f31 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.009671] env[63355]: DEBUG nova.compute.manager [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1174.195715] env[63355]: DEBUG nova.compute.utils [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1174.258061] env[63355]: DEBUG oslo_concurrency.lockutils [None req-a3ec1af3-3bf5-4cb3-a601-2102fc1b6294 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "8a1a2c12-5042-43fd-bb62-0ccbd0f9eb50" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.226s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.401730] env[63355]: INFO nova.compute.manager [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Took 13.90 seconds to build instance. [ 1174.699582] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.903628] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2b5f7d0c-a11a-47d9-bfaa-685f4a1de164 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "ddaab84c-5f1d-4122-9408-2043f30a2227" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.407s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.019160] env[63355]: DEBUG nova.compute.manager [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1175.039704] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1175.040013] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1175.040191] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1175.040378] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1175.040530] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1175.040683] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1175.040893] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1175.041235] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1175.041455] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1175.042015] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1175.042015] env[63355]: DEBUG nova.virt.hardware [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1175.042786] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74816fc8-c6cb-4d63-bc34-10b36096fe9b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.051544] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c8bfae-deb8-4eed-935a-aa33f778a13b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.150030] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "500cef8e-f058-4597-8410-ff83bfab06eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.151339] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "500cef8e-f058-4597-8410-ff83bfab06eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.153130] env[63355]: DEBUG nova.compute.manager [req-b130f663-f441-463b-abef-f4c0966200b6 req-528833d5-f2e9-4f76-ba77-b59f3d6c0134 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Received event network-changed-bba4c0ca-7f31-49ce-92ec-703f913914ce {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1175.153772] env[63355]: DEBUG nova.compute.manager [req-b130f663-f441-463b-abef-f4c0966200b6 req-528833d5-f2e9-4f76-ba77-b59f3d6c0134 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Refreshing instance network info cache due to event network-changed-bba4c0ca-7f31-49ce-92ec-703f913914ce. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1175.153986] env[63355]: DEBUG oslo_concurrency.lockutils [req-b130f663-f441-463b-abef-f4c0966200b6 req-528833d5-f2e9-4f76-ba77-b59f3d6c0134 service nova] Acquiring lock "refresh_cache-ddaab84c-5f1d-4122-9408-2043f30a2227" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.154154] env[63355]: DEBUG oslo_concurrency.lockutils [req-b130f663-f441-463b-abef-f4c0966200b6 req-528833d5-f2e9-4f76-ba77-b59f3d6c0134 service nova] Acquired lock "refresh_cache-ddaab84c-5f1d-4122-9408-2043f30a2227" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.154323] env[63355]: DEBUG nova.network.neutron [req-b130f663-f441-463b-abef-f4c0966200b6 req-528833d5-f2e9-4f76-ba77-b59f3d6c0134 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Refreshing network info cache for port bba4c0ca-7f31-49ce-92ec-703f913914ce {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1175.325449] env[63355]: DEBUG nova.network.neutron [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Successfully updated port: 417738e8-effd-45de-9c4b-b0eda6d5ccfc {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1175.657156] env[63355]: DEBUG nova.compute.manager [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1175.767729] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.767987] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.768232] env[63355]: INFO nova.compute.manager [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Attaching volume 48dd3241-16cb-4df7-9a85-da673593e136 to /dev/sdb [ 1175.801507] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4d68ba-df77-44b3-b8d8-b99218e67125 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.809237] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff6488d-06cb-4b21-a5bc-95ddba0c40d5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.822341] env[63355]: DEBUG nova.virt.block_device [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Updating existing volume attachment record: e8fe145f-c2d4-4f99-83e0-737d4be83a0b {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1175.828212] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "refresh_cache-1c054e10-f650-4b47-a233-3bf17482a9b8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.828356] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "refresh_cache-1c054e10-f650-4b47-a233-3bf17482a9b8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.828507] env[63355]: DEBUG nova.network.neutron [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1175.901820] env[63355]: DEBUG nova.network.neutron [req-b130f663-f441-463b-abef-f4c0966200b6 req-528833d5-f2e9-4f76-ba77-b59f3d6c0134 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Updated VIF entry in instance network info cache for port bba4c0ca-7f31-49ce-92ec-703f913914ce. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1175.902233] env[63355]: DEBUG nova.network.neutron [req-b130f663-f441-463b-abef-f4c0966200b6 req-528833d5-f2e9-4f76-ba77-b59f3d6c0134 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Updating instance_info_cache with network_info: [{"id": "bba4c0ca-7f31-49ce-92ec-703f913914ce", "address": "fa:16:3e:c0:a0:5a", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba4c0ca-7f", "ovs_interfaceid": "bba4c0ca-7f31-49ce-92ec-703f913914ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.183138] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.183477] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.185515] env[63355]: INFO nova.compute.claims [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1176.359879] env[63355]: DEBUG nova.network.neutron [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1176.404942] env[63355]: DEBUG oslo_concurrency.lockutils [req-b130f663-f441-463b-abef-f4c0966200b6 req-528833d5-f2e9-4f76-ba77-b59f3d6c0134 service nova] Releasing lock "refresh_cache-ddaab84c-5f1d-4122-9408-2043f30a2227" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.499638] env[63355]: DEBUG nova.network.neutron [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Updating instance_info_cache with network_info: [{"id": "417738e8-effd-45de-9c4b-b0eda6d5ccfc", "address": "fa:16:3e:b6:c1:1d", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap417738e8-ef", "ovs_interfaceid": "417738e8-effd-45de-9c4b-b0eda6d5ccfc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.002516] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "refresh_cache-1c054e10-f650-4b47-a233-3bf17482a9b8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.002930] env[63355]: DEBUG nova.compute.manager [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Instance network_info: |[{"id": "417738e8-effd-45de-9c4b-b0eda6d5ccfc", "address": "fa:16:3e:b6:c1:1d", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap417738e8-ef", "ovs_interfaceid": "417738e8-effd-45de-9c4b-b0eda6d5ccfc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1177.003415] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:c1:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '417738e8-effd-45de-9c4b-b0eda6d5ccfc', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1177.010924] env[63355]: DEBUG oslo.service.loopingcall [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1177.011159] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1177.011387] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9d4f96d-2a22-4d1e-b089-230f6fba3da9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.031331] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1177.031331] env[63355]: value = "task-1350235" [ 1177.031331] env[63355]: _type = "Task" [ 1177.031331] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.038742] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350235, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.181524] env[63355]: DEBUG nova.compute.manager [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Received event network-vif-plugged-417738e8-effd-45de-9c4b-b0eda6d5ccfc {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1177.181758] env[63355]: DEBUG oslo_concurrency.lockutils [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] Acquiring lock "1c054e10-f650-4b47-a233-3bf17482a9b8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.181973] env[63355]: DEBUG oslo_concurrency.lockutils [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.182167] env[63355]: DEBUG oslo_concurrency.lockutils [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.182343] env[63355]: DEBUG nova.compute.manager [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] No waiting events found dispatching network-vif-plugged-417738e8-effd-45de-9c4b-b0eda6d5ccfc {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1177.182533] env[63355]: WARNING nova.compute.manager [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Received unexpected event network-vif-plugged-417738e8-effd-45de-9c4b-b0eda6d5ccfc for instance with vm_state building and task_state spawning. [ 1177.182742] env[63355]: DEBUG nova.compute.manager [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Received event network-changed-417738e8-effd-45de-9c4b-b0eda6d5ccfc {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1177.182908] env[63355]: DEBUG nova.compute.manager [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Refreshing instance network info cache due to event network-changed-417738e8-effd-45de-9c4b-b0eda6d5ccfc. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1177.183111] env[63355]: DEBUG oslo_concurrency.lockutils [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] Acquiring lock "refresh_cache-1c054e10-f650-4b47-a233-3bf17482a9b8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.183254] env[63355]: DEBUG oslo_concurrency.lockutils [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] Acquired lock "refresh_cache-1c054e10-f650-4b47-a233-3bf17482a9b8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.183413] env[63355]: DEBUG nova.network.neutron [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Refreshing network info cache for port 417738e8-effd-45de-9c4b-b0eda6d5ccfc {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1177.269246] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3462c85e-2e8d-4231-9be8-a5b68b7c5b4d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.276947] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6baceba3-f924-4d5e-bc62-2a205214c46b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.306201] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4cd284-c65d-477c-890a-c526e1309984 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.313376] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f808db5c-6403-4a66-9001-ce4b3fca944d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.327116] env[63355]: DEBUG nova.compute.provider_tree [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.541402] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350235, 'name': CreateVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.830972] env[63355]: DEBUG nova.scheduler.client.report [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1177.894467] env[63355]: DEBUG nova.network.neutron [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Updated VIF entry in instance network info cache for port 417738e8-effd-45de-9c4b-b0eda6d5ccfc. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1177.894801] env[63355]: DEBUG nova.network.neutron [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Updating instance_info_cache with network_info: [{"id": "417738e8-effd-45de-9c4b-b0eda6d5ccfc", "address": "fa:16:3e:b6:c1:1d", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap417738e8-ef", "ovs_interfaceid": "417738e8-effd-45de-9c4b-b0eda6d5ccfc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.041326] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350235, 'name': CreateVM_Task, 'duration_secs': 0.593994} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.041589] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1178.042191] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.042369] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.042697] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1178.042956] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89e32c4d-2130-40be-920d-e4e9ed118d76 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.047236] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1178.047236] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52812d8f-ebcc-6c95-91b4-f6a77c529be0" [ 1178.047236] env[63355]: _type = "Task" [ 1178.047236] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.054640] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52812d8f-ebcc-6c95-91b4-f6a77c529be0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.338130] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.154s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.338817] env[63355]: DEBUG nova.compute.manager [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1178.397443] env[63355]: DEBUG oslo_concurrency.lockutils [req-2e30ae83-df07-4530-8115-330cbff08231 req-f0a6e724-8293-4d63-8cc8-31c672229cd3 service nova] Releasing lock "refresh_cache-1c054e10-f650-4b47-a233-3bf17482a9b8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.557526] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52812d8f-ebcc-6c95-91b4-f6a77c529be0, 'name': SearchDatastore_Task, 'duration_secs': 0.0104} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.558146] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.558146] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1178.558309] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.558908] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.558908] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1178.558908] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-67790bec-dda9-465c-b976-22b31b992821 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.566442] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1178.566621] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1178.567288] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b0497b0-5148-41ab-a502-e7ad77851054 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.571818] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1178.571818] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5293c1f3-112e-a8e2-f773-f9b6e38380fe" [ 1178.571818] env[63355]: _type = "Task" [ 1178.571818] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.578848] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5293c1f3-112e-a8e2-f773-f9b6e38380fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.844222] env[63355]: DEBUG nova.compute.utils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1178.845642] env[63355]: DEBUG nova.compute.manager [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1178.845813] env[63355]: DEBUG nova.network.neutron [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1178.883867] env[63355]: DEBUG nova.policy [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54ad2682257b4250a8a96f4ba6a9523b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b3fee9bc99d49ea9de53d5dce52c79d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1179.082264] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5293c1f3-112e-a8e2-f773-f9b6e38380fe, 'name': SearchDatastore_Task, 'duration_secs': 0.007466} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.082728] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e795d10-f0ba-4118-a4a6-a3b42231e9f6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.087794] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1179.087794] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52bd5d53-5be6-af9a-0a9d-0a7ce7ff4a64" [ 1179.087794] env[63355]: _type = "Task" [ 1179.087794] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.095284] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52bd5d53-5be6-af9a-0a9d-0a7ce7ff4a64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.138419] env[63355]: DEBUG nova.network.neutron [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Successfully created port: 57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1179.349281] env[63355]: DEBUG nova.compute.manager [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1179.598055] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52bd5d53-5be6-af9a-0a9d-0a7ce7ff4a64, 'name': SearchDatastore_Task, 'duration_secs': 0.00874} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.598055] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.598349] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 1c054e10-f650-4b47-a233-3bf17482a9b8/1c054e10-f650-4b47-a233-3bf17482a9b8.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1179.598617] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0235b861-9507-41b9-b60d-3bb66a24c550 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.605116] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1179.605116] env[63355]: value = "task-1350237" [ 1179.605116] env[63355]: _type = "Task" [ 1179.605116] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.612125] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.115952] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350237, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442898} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.116338] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 1c054e10-f650-4b47-a233-3bf17482a9b8/1c054e10-f650-4b47-a233-3bf17482a9b8.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1180.116606] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1180.116907] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4131d36b-e211-4a7a-892d-caffce01b804 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.123427] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1180.123427] env[63355]: value = "task-1350238" [ 1180.123427] env[63355]: _type = "Task" [ 1180.123427] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.130646] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350238, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.359855] env[63355]: DEBUG nova.compute.manager [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1180.370104] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1180.372055] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287769', 'volume_id': '48dd3241-16cb-4df7-9a85-da673593e136', 'name': 'volume-48dd3241-16cb-4df7-9a85-da673593e136', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d332e8f-4c89-4050-b5b7-3cd034f9edfe', 'attached_at': '', 'detached_at': '', 'volume_id': '48dd3241-16cb-4df7-9a85-da673593e136', 'serial': '48dd3241-16cb-4df7-9a85-da673593e136'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1180.372215] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86d6bb4-8a02-42b8-9e8c-035cd809caac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.389281] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c143dd-5ddd-45d2-89db-8914f099206d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.393927] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1180.394170] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1180.394332] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1180.394518] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1180.394670] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1180.394820] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1180.395040] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1180.395210] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1180.395377] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1180.395552] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1180.395740] env[63355]: DEBUG nova.virt.hardware [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1180.396477] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad11561-a27e-4ba7-8e00-1e7043d3d676 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.418158] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] volume-48dd3241-16cb-4df7-9a85-da673593e136/volume-48dd3241-16cb-4df7-9a85-da673593e136.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1180.418757] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b6723a2-d1ba-4d40-8057-c7dacd0447c9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.434409] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee21de7-af2e-4127-912e-26caa0a8afed {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.439053] env[63355]: DEBUG oslo_vmware.api [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1180.439053] env[63355]: value = "task-1350239" [ 1180.439053] env[63355]: _type = "Task" [ 1180.439053] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.454831] env[63355]: DEBUG oslo_vmware.api [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350239, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.608417] env[63355]: DEBUG nova.compute.manager [req-579f3c55-bcfc-4450-a766-7d4ac23762fd req-f0801dcc-b26d-4cca-8063-421ad6330f06 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received event network-vif-plugged-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1180.608681] env[63355]: DEBUG oslo_concurrency.lockutils [req-579f3c55-bcfc-4450-a766-7d4ac23762fd req-f0801dcc-b26d-4cca-8063-421ad6330f06 service nova] Acquiring lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.608907] env[63355]: DEBUG oslo_concurrency.lockutils [req-579f3c55-bcfc-4450-a766-7d4ac23762fd req-f0801dcc-b26d-4cca-8063-421ad6330f06 service nova] Lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.609099] env[63355]: DEBUG oslo_concurrency.lockutils [req-579f3c55-bcfc-4450-a766-7d4ac23762fd req-f0801dcc-b26d-4cca-8063-421ad6330f06 service nova] Lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.609275] env[63355]: DEBUG nova.compute.manager [req-579f3c55-bcfc-4450-a766-7d4ac23762fd req-f0801dcc-b26d-4cca-8063-421ad6330f06 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] No waiting events found dispatching network-vif-plugged-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1180.609443] env[63355]: WARNING nova.compute.manager [req-579f3c55-bcfc-4450-a766-7d4ac23762fd req-f0801dcc-b26d-4cca-8063-421ad6330f06 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received unexpected event network-vif-plugged-57106266-129a-4058-8690-6deccdf8ab52 for instance with vm_state building and task_state spawning. [ 1180.636062] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350238, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059022} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.636571] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1180.637584] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79d3428-418a-4a8b-afd7-5f550ff2330c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.660310] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 1c054e10-f650-4b47-a233-3bf17482a9b8/1c054e10-f650-4b47-a233-3bf17482a9b8.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1180.660620] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38dc3274-3f1e-418d-9864-1988f22ad2c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.675969] env[63355]: DEBUG nova.network.neutron [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Successfully updated port: 57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1180.683152] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1180.683152] env[63355]: value = "task-1350240" [ 1180.683152] env[63355]: _type = "Task" [ 1180.683152] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.688928] env[63355]: DEBUG nova.compute.manager [req-8e0233bf-525b-4cf1-86ae-8afc9c4fd56c req-8fdd0d98-74c0-410f-a034-1eb757df7e48 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received event network-changed-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1180.689144] env[63355]: DEBUG nova.compute.manager [req-8e0233bf-525b-4cf1-86ae-8afc9c4fd56c req-8fdd0d98-74c0-410f-a034-1eb757df7e48 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Refreshing instance network info cache due to event network-changed-57106266-129a-4058-8690-6deccdf8ab52. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1180.689360] env[63355]: DEBUG oslo_concurrency.lockutils [req-8e0233bf-525b-4cf1-86ae-8afc9c4fd56c req-8fdd0d98-74c0-410f-a034-1eb757df7e48 service nova] Acquiring lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.689509] env[63355]: DEBUG oslo_concurrency.lockutils [req-8e0233bf-525b-4cf1-86ae-8afc9c4fd56c req-8fdd0d98-74c0-410f-a034-1eb757df7e48 service nova] Acquired lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.689676] env[63355]: DEBUG nova.network.neutron [req-8e0233bf-525b-4cf1-86ae-8afc9c4fd56c req-8fdd0d98-74c0-410f-a034-1eb757df7e48 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Refreshing network info cache for port 57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1180.697336] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350240, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.951550] env[63355]: DEBUG oslo_vmware.api [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.189021] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.196481] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350240, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.224821] env[63355]: DEBUG nova.network.neutron [req-8e0233bf-525b-4cf1-86ae-8afc9c4fd56c req-8fdd0d98-74c0-410f-a034-1eb757df7e48 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1181.302034] env[63355]: DEBUG nova.network.neutron [req-8e0233bf-525b-4cf1-86ae-8afc9c4fd56c req-8fdd0d98-74c0-410f-a034-1eb757df7e48 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.451702] env[63355]: DEBUG oslo_vmware.api [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350239, 'name': ReconfigVM_Task, 'duration_secs': 0.709173} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.451965] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Reconfigured VM instance instance-00000068 to attach disk [datastore1] volume-48dd3241-16cb-4df7-9a85-da673593e136/volume-48dd3241-16cb-4df7-9a85-da673593e136.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1181.456634] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0929935-0641-49ee-94aa-3ae238a4174e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.472452] env[63355]: DEBUG oslo_vmware.api [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1181.472452] env[63355]: value = "task-1350241" [ 1181.472452] env[63355]: _type = "Task" [ 1181.472452] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.480746] env[63355]: DEBUG oslo_vmware.api [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350241, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.694425] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350240, 'name': ReconfigVM_Task, 'duration_secs': 0.970281} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.694709] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 1c054e10-f650-4b47-a233-3bf17482a9b8/1c054e10-f650-4b47-a233-3bf17482a9b8.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1181.695364] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46e62a6e-f9b4-4076-bd77-a0b1bdeefce1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.701671] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1181.701671] env[63355]: value = "task-1350242" [ 1181.701671] env[63355]: _type = "Task" [ 1181.701671] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.710417] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350242, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.805076] env[63355]: DEBUG oslo_concurrency.lockutils [req-8e0233bf-525b-4cf1-86ae-8afc9c4fd56c req-8fdd0d98-74c0-410f-a034-1eb757df7e48 service nova] Releasing lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.805476] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.805643] env[63355]: DEBUG nova.network.neutron [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1181.982802] env[63355]: DEBUG oslo_vmware.api [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350241, 'name': ReconfigVM_Task, 'duration_secs': 0.246217} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.983099] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287769', 'volume_id': '48dd3241-16cb-4df7-9a85-da673593e136', 'name': 'volume-48dd3241-16cb-4df7-9a85-da673593e136', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d332e8f-4c89-4050-b5b7-3cd034f9edfe', 'attached_at': '', 'detached_at': '', 'volume_id': '48dd3241-16cb-4df7-9a85-da673593e136', 'serial': '48dd3241-16cb-4df7-9a85-da673593e136'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1182.212126] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350242, 'name': Rename_Task, 'duration_secs': 0.479147} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.212476] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1182.212652] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f5f9788-0908-44ba-9da5-9b1195163be8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.218783] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1182.218783] env[63355]: value = "task-1350243" [ 1182.218783] env[63355]: _type = "Task" [ 1182.218783] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.228008] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350243, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.336501] env[63355]: DEBUG nova.network.neutron [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1182.465621] env[63355]: DEBUG nova.network.neutron [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updating instance_info_cache with network_info: [{"id": "57106266-129a-4058-8690-6deccdf8ab52", "address": "fa:16:3e:9e:a7:c8", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57106266-12", "ovs_interfaceid": "57106266-129a-4058-8690-6deccdf8ab52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.728447] env[63355]: DEBUG oslo_vmware.api [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350243, 'name': PowerOnVM_Task, 'duration_secs': 0.461246} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.728822] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1182.729075] env[63355]: INFO nova.compute.manager [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Took 7.71 seconds to spawn the instance on the hypervisor. [ 1182.729291] env[63355]: DEBUG nova.compute.manager [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1182.730110] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b2f1a8-eee7-4c7e-bc8a-d842d639dee8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.968861] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.969177] env[63355]: DEBUG nova.compute.manager [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Instance network_info: |[{"id": "57106266-129a-4058-8690-6deccdf8ab52", "address": "fa:16:3e:9e:a7:c8", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57106266-12", "ovs_interfaceid": "57106266-129a-4058-8690-6deccdf8ab52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1182.969659] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:a7:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57106266-129a-4058-8690-6deccdf8ab52', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1182.977183] env[63355]: DEBUG oslo.service.loopingcall [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1182.977401] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1182.977623] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eddf821c-037d-4766-a0d3-80f456476ab1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.997552] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1182.997552] env[63355]: value = "task-1350244" [ 1182.997552] env[63355]: _type = "Task" [ 1182.997552] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.004728] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350244, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.018668] env[63355]: DEBUG nova.objects.instance [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'flavor' on Instance uuid 8d332e8f-4c89-4050-b5b7-3cd034f9edfe {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1183.246752] env[63355]: INFO nova.compute.manager [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Took 12.45 seconds to build instance. [ 1183.507022] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350244, 'name': CreateVM_Task, 'duration_secs': 0.345301} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.508021] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1183.508147] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.508329] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.508691] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1183.508946] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1204c7a3-df65-47aa-9d36-8c1518dc1bbb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.513280] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1183.513280] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527b2c8a-87d9-07f1-1fd0-679059fa1660" [ 1183.513280] env[63355]: _type = "Task" [ 1183.513280] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.520773] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527b2c8a-87d9-07f1-1fd0-679059fa1660, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.523156] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0f245f77-c963-4740-800a-4d7db516edc6 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.755s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.748701] env[63355]: DEBUG oslo_concurrency.lockutils [None req-b59beea1-2664-4dd1-9e86-9e6802dacab2 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.956s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.024113] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527b2c8a-87d9-07f1-1fd0-679059fa1660, 'name': SearchDatastore_Task, 'duration_secs': 0.009929} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.024113] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.024339] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1184.024454] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1184.024607] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.024789] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1184.025284] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00f588fb-8bb0-4b0e-9c64-6dee951e9de8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.032770] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1184.032943] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1184.033610] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b69caa35-fcef-4251-b47a-9c619cb323e7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.039043] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1184.039043] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a2465a-91a2-ba08-ea54-3568e6d7106e" [ 1184.039043] env[63355]: _type = "Task" [ 1184.039043] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.045500] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a2465a-91a2-ba08-ea54-3568e6d7106e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.288094] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "1c054e10-f650-4b47-a233-3bf17482a9b8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.288454] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.288454] env[63355]: DEBUG nova.compute.manager [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.289326] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a91bf7-6088-41b1-97a3-8b3b5d36fd32 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.295699] env[63355]: DEBUG nova.compute.manager [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63355) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1184.296254] env[63355]: DEBUG nova.objects.instance [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lazy-loading 'flavor' on Instance uuid 1c054e10-f650-4b47-a233-3bf17482a9b8 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1184.548847] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a2465a-91a2-ba08-ea54-3568e6d7106e, 'name': SearchDatastore_Task, 'duration_secs': 0.007986} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.549587] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1238c4d2-49d8-43f7-81d8-e211675f5f32 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.554405] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1184.554405] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52539856-25fb-c0e5-74fb-cd704ff71c9d" [ 1184.554405] env[63355]: _type = "Task" [ 1184.554405] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.561607] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52539856-25fb-c0e5-74fb-cd704ff71c9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.801603] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1184.801774] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-22b70916-01dd-4fa3-be68-c94294371517 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.809113] env[63355]: DEBUG oslo_vmware.api [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1184.809113] env[63355]: value = "task-1350245" [ 1184.809113] env[63355]: _type = "Task" [ 1184.809113] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.817144] env[63355]: DEBUG oslo_vmware.api [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350245, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.879494] env[63355]: DEBUG oslo_concurrency.lockutils [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.879744] env[63355]: DEBUG oslo_concurrency.lockutils [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.064486] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52539856-25fb-c0e5-74fb-cd704ff71c9d, 'name': SearchDatastore_Task, 'duration_secs': 0.009792} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.064637] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.064884] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 500cef8e-f058-4597-8410-ff83bfab06eb/500cef8e-f058-4597-8410-ff83bfab06eb.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1185.065160] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-900174f7-30a1-4772-b1bb-9aa8b1a354f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.071413] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1185.071413] env[63355]: value = "task-1350246" [ 1185.071413] env[63355]: _type = "Task" [ 1185.071413] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.078254] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350246, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.318995] env[63355]: DEBUG oslo_vmware.api [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350245, 'name': PowerOffVM_Task, 'duration_secs': 0.186606} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.319368] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1185.319489] env[63355]: DEBUG nova.compute.manager [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1185.320253] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38570db5-d43f-46c1-9ca9-a11ec383a3f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.383446] env[63355]: DEBUG nova.compute.utils [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1185.580740] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350246, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455699} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.582036] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 500cef8e-f058-4597-8410-ff83bfab06eb/500cef8e-f058-4597-8410-ff83bfab06eb.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1185.582036] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1185.582036] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f531906-98a9-494f-8fb3-04e7c20d89e8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.588348] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1185.588348] env[63355]: value = "task-1350247" [ 1185.588348] env[63355]: _type = "Task" [ 1185.588348] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.595396] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350247, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.831901] env[63355]: DEBUG oslo_concurrency.lockutils [None req-9d8016ac-ec17-4f4a-ae46-bd45eb68218a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.886219] env[63355]: DEBUG oslo_concurrency.lockutils [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.100020] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350247, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059328} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.100304] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1186.101057] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367b07a4-d5ee-4a63-b0ca-c5e7d4da3f53 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.122878] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 500cef8e-f058-4597-8410-ff83bfab06eb/500cef8e-f058-4597-8410-ff83bfab06eb.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1186.123160] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79647102-dbb0-4b1b-8b9e-2a8a9ca3c54a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.142153] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1186.142153] env[63355]: value = "task-1350248" [ 1186.142153] env[63355]: _type = "Task" [ 1186.142153] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.152711] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350248, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.652136] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350248, 'name': ReconfigVM_Task, 'duration_secs': 0.314878} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.652543] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 500cef8e-f058-4597-8410-ff83bfab06eb/500cef8e-f058-4597-8410-ff83bfab06eb.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1186.652991] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ad61ae7-b181-4311-b6a9-63f8bf3de0d8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.658909] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1186.658909] env[63355]: value = "task-1350249" [ 1186.658909] env[63355]: _type = "Task" [ 1186.658909] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.665854] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350249, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.709022] env[63355]: INFO nova.compute.manager [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Rebuilding instance [ 1186.747267] env[63355]: DEBUG nova.compute.manager [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1186.748176] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9411cfa-aaf1-494a-b360-d8ee8dd5c272 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.945209] env[63355]: DEBUG oslo_concurrency.lockutils [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.945487] env[63355]: DEBUG oslo_concurrency.lockutils [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.945764] env[63355]: INFO nova.compute.manager [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Attaching volume b65b95db-ab9c-4ae3-b295-22a877310891 to /dev/sdc [ 1186.975616] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb96b05e-e563-47c3-b645-233ae4ae2870 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.982463] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d9a431-c895-4334-92af-d272c76f5ecd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.996014] env[63355]: DEBUG nova.virt.block_device [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Updating existing volume attachment record: 7b7a6732-ef7e-4b72-9af8-5e6186b99d55 {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1187.167789] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350249, 'name': Rename_Task, 'duration_secs': 0.133239} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.168106] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1187.168354] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-838b2e71-3bf8-49ff-8c09-1770692a1ac7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.173786] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1187.173786] env[63355]: value = "task-1350251" [ 1187.173786] env[63355]: _type = "Task" [ 1187.173786] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.182378] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350251, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.260434] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1187.260697] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f24c237c-160f-4286-a7d3-52cce5ea178a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.267469] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1187.267469] env[63355]: value = "task-1350252" [ 1187.267469] env[63355]: _type = "Task" [ 1187.267469] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.275254] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350252, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.683605] env[63355]: DEBUG oslo_vmware.api [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350251, 'name': PowerOnVM_Task, 'duration_secs': 0.426384} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.683988] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1187.684182] env[63355]: INFO nova.compute.manager [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Took 7.32 seconds to spawn the instance on the hypervisor. [ 1187.684371] env[63355]: DEBUG nova.compute.manager [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1187.685142] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf88c26-0cc6-40ce-9397-230eec916539 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.777668] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] VM already powered off {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1187.777919] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1187.778703] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f269c55-4da0-476d-b790-be33111fb228 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.785095] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1187.785337] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1cb8d74-d930-4811-81c7-1a99977f9198 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.849971] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1187.850148] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1187.850338] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleting the datastore file [datastore1] 1c054e10-f650-4b47-a233-3bf17482a9b8 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.850611] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7fc971d0-b3af-4bc0-b518-c19b094d8677 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.857148] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1187.857148] env[63355]: value = "task-1350254" [ 1187.857148] env[63355]: _type = "Task" [ 1187.857148] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.865339] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350254, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.200538] env[63355]: INFO nova.compute.manager [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Took 12.04 seconds to build instance. [ 1188.366880] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350254, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213026} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.367659] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1188.367659] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1188.367659] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1188.489778] env[63355]: DEBUG nova.compute.manager [req-3c2d8acc-d895-414c-b017-2638c5d74706 req-e99f5afe-f05b-4680-bf94-33ae051a5827 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received event network-changed-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1188.489993] env[63355]: DEBUG nova.compute.manager [req-3c2d8acc-d895-414c-b017-2638c5d74706 req-e99f5afe-f05b-4680-bf94-33ae051a5827 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Refreshing instance network info cache due to event network-changed-57106266-129a-4058-8690-6deccdf8ab52. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1188.490227] env[63355]: DEBUG oslo_concurrency.lockutils [req-3c2d8acc-d895-414c-b017-2638c5d74706 req-e99f5afe-f05b-4680-bf94-33ae051a5827 service nova] Acquiring lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1188.490368] env[63355]: DEBUG oslo_concurrency.lockutils [req-3c2d8acc-d895-414c-b017-2638c5d74706 req-e99f5afe-f05b-4680-bf94-33ae051a5827 service nova] Acquired lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.490528] env[63355]: DEBUG nova.network.neutron [req-3c2d8acc-d895-414c-b017-2638c5d74706 req-e99f5afe-f05b-4680-bf94-33ae051a5827 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Refreshing network info cache for port 57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1188.703297] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cdc964c3-9a3e-48f7-9926-870a2ca9c391 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "500cef8e-f058-4597-8410-ff83bfab06eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.553s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.132763] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.197328] env[63355]: DEBUG nova.network.neutron [req-3c2d8acc-d895-414c-b017-2638c5d74706 req-e99f5afe-f05b-4680-bf94-33ae051a5827 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updated VIF entry in instance network info cache for port 57106266-129a-4058-8690-6deccdf8ab52. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1189.197685] env[63355]: DEBUG nova.network.neutron [req-3c2d8acc-d895-414c-b017-2638c5d74706 req-e99f5afe-f05b-4680-bf94-33ae051a5827 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updating instance_info_cache with network_info: [{"id": "57106266-129a-4058-8690-6deccdf8ab52", "address": "fa:16:3e:9e:a7:c8", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57106266-12", "ovs_interfaceid": "57106266-129a-4058-8690-6deccdf8ab52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.403355] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1189.403610] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1189.403768] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1189.403954] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1189.404121] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1189.404273] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1189.404479] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1189.404638] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1189.404804] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1189.404968] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1189.405154] env[63355]: DEBUG nova.virt.hardware [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1189.406008] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f190b3b0-d5d1-4d63-989f-2953d4eb66b1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.413338] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6492a95-4dd4-4fa2-bd36-4af66b4cb048 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.426338] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:c1:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '417738e8-effd-45de-9c4b-b0eda6d5ccfc', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1189.433417] env[63355]: DEBUG oslo.service.loopingcall [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1189.433637] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1189.433832] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0fc8d3d3-d8ad-43e5-9f29-00005a6c7558 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.451330] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1189.451330] env[63355]: value = "task-1350256" [ 1189.451330] env[63355]: _type = "Task" [ 1189.451330] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.458095] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350256, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.636574] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.636829] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.637093] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.637303] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1189.638362] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6897dbd5-31cd-4a31-aec9-918b870584c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.647868] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a79154-af7e-4cae-8130-9c25fb620c8d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.661709] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e964458c-6c78-4cf8-8752-23cb2d2c0e77 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.667975] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f12fc1-7784-4a96-aa04-d14dc5f87084 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.695602] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180929MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1189.695749] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.695935] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.700283] env[63355]: DEBUG oslo_concurrency.lockutils [req-3c2d8acc-d895-414c-b017-2638c5d74706 req-e99f5afe-f05b-4680-bf94-33ae051a5827 service nova] Releasing lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1189.963200] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350256, 'name': CreateVM_Task, 'duration_secs': 0.28848} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.963562] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1189.964032] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.964210] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.964566] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1189.964926] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9700941-4619-4e34-a73c-b7b2e9a3ee6b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.969116] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1189.969116] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b12dce-f7e1-9ea6-109b-52bcb2a0e5b3" [ 1189.969116] env[63355]: _type = "Task" [ 1189.969116] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.976381] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b12dce-f7e1-9ea6-109b-52bcb2a0e5b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.478904] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52b12dce-f7e1-9ea6-109b-52bcb2a0e5b3, 'name': SearchDatastore_Task, 'duration_secs': 0.00883} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.479254] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.479491] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1190.479750] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.479915] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.480118] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1190.480390] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe8ad297-f80c-4582-b972-b42b9f7d72c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.487968] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1190.488163] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1190.488884] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b2271b7-eed7-4d53-99a1-a64461db397a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.493839] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1190.493839] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e4a8a4-8539-0c8c-b1a6-5622611ccbda" [ 1190.493839] env[63355]: _type = "Task" [ 1190.493839] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.501287] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e4a8a4-8539-0c8c-b1a6-5622611ccbda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.722025] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1190.722025] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8db23054-4556-496a-a4b1-a24c71c30f50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1190.722025] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8d332e8f-4c89-4050-b5b7-3cd034f9edfe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1190.722025] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance ddaab84c-5f1d-4122-9408-2043f30a2227 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1190.722319] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 1c054e10-f650-4b47-a233-3bf17482a9b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1190.722319] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 500cef8e-f058-4597-8410-ff83bfab06eb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1190.722380] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1190.722485] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1190.801127] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f77a0e-28b7-4302-a236-150868c2a000 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.808462] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83684198-c047-4e7d-b25d-76880a5dcc4c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.837350] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621dd4e1-1ef1-4838-93d4-b4d7b13abcc5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.844746] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c825ce-74d9-42b7-8094-14fc2aed1b6b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.857644] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1191.006096] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52e4a8a4-8539-0c8c-b1a6-5622611ccbda, 'name': SearchDatastore_Task, 'duration_secs': 0.008264} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.006742] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b144196-06d0-4694-a78f-82c201b2a8e6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.011507] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1191.011507] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52eb17e0-257f-56cf-3cdf-4bf283ceaac9" [ 1191.011507] env[63355]: _type = "Task" [ 1191.011507] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.018526] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52eb17e0-257f-56cf-3cdf-4bf283ceaac9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.360688] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1191.522021] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52eb17e0-257f-56cf-3cdf-4bf283ceaac9, 'name': SearchDatastore_Task, 'duration_secs': 0.008555} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.522303] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.522564] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 1c054e10-f650-4b47-a233-3bf17482a9b8/1c054e10-f650-4b47-a233-3bf17482a9b8.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1191.522849] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e835fa74-5b7b-4145-86c7-2465357b5f12 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.529377] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1191.529377] env[63355]: value = "task-1350257" [ 1191.529377] env[63355]: _type = "Task" [ 1191.529377] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.536556] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350257, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.537483] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1191.537707] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287772', 'volume_id': 'b65b95db-ab9c-4ae3-b295-22a877310891', 'name': 'volume-b65b95db-ab9c-4ae3-b295-22a877310891', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d332e8f-4c89-4050-b5b7-3cd034f9edfe', 'attached_at': '', 'detached_at': '', 'volume_id': 'b65b95db-ab9c-4ae3-b295-22a877310891', 'serial': 'b65b95db-ab9c-4ae3-b295-22a877310891'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1191.538450] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12899d9f-7585-4fbf-865f-45ef589ebb15 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.552753] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c55f2c3-0e22-4dbd-8524-ddd8262fe70d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.578326] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] volume-b65b95db-ab9c-4ae3-b295-22a877310891/volume-b65b95db-ab9c-4ae3-b295-22a877310891.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1191.578564] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2a5e024-9a8e-4217-9dff-e376d4575ae6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.596280] env[63355]: DEBUG oslo_vmware.api [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1191.596280] env[63355]: value = "task-1350258" [ 1191.596280] env[63355]: _type = "Task" [ 1191.596280] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.604891] env[63355]: DEBUG oslo_vmware.api [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350258, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.865726] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1191.865983] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.170s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.039013] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350257, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434388} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.039489] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 1c054e10-f650-4b47-a233-3bf17482a9b8/1c054e10-f650-4b47-a233-3bf17482a9b8.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1192.039489] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1192.039748] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf3bb8f0-69ef-4b2c-ae0b-a924f597aafa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.045956] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1192.045956] env[63355]: value = "task-1350259" [ 1192.045956] env[63355]: _type = "Task" [ 1192.045956] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.052792] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350259, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.105706] env[63355]: DEBUG oslo_vmware.api [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350258, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.555148] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350259, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063563} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.555426] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1192.556148] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e397082-0970-4b5d-bf78-ce6e5368b2a0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.577826] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 1c054e10-f650-4b47-a233-3bf17482a9b8/1c054e10-f650-4b47-a233-3bf17482a9b8.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1192.578065] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-992ac518-a680-4448-8df7-7c21c1131d0f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.595955] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1192.595955] env[63355]: value = "task-1350260" [ 1192.595955] env[63355]: _type = "Task" [ 1192.595955] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.606949] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350260, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.609837] env[63355]: DEBUG oslo_vmware.api [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350258, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.108900] env[63355]: DEBUG oslo_vmware.api [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350258, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.111820] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350260, 'name': ReconfigVM_Task, 'duration_secs': 0.272066} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.112098] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 1c054e10-f650-4b47-a233-3bf17482a9b8/1c054e10-f650-4b47-a233-3bf17482a9b8.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1193.112692] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddbafc2e-62e2-4f3f-a24e-406009928af8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.118849] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1193.118849] env[63355]: value = "task-1350261" [ 1193.118849] env[63355]: _type = "Task" [ 1193.118849] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.125873] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350261, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.609391] env[63355]: DEBUG oslo_vmware.api [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350258, 'name': ReconfigVM_Task, 'duration_secs': 1.533257} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.609717] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Reconfigured VM instance instance-00000068 to attach disk [datastore2] volume-b65b95db-ab9c-4ae3-b295-22a877310891/volume-b65b95db-ab9c-4ae3-b295-22a877310891.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1193.614288] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f0189b5-ce44-4298-ba40-0a75a063189b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.632187] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350261, 'name': Rename_Task, 'duration_secs': 0.128719} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.633335] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1193.633632] env[63355]: DEBUG oslo_vmware.api [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1193.633632] env[63355]: value = "task-1350262" [ 1193.633632] env[63355]: _type = "Task" [ 1193.633632] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.633816] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2693f94-1ea3-49a6-aaec-cf85d8c26651 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.642293] env[63355]: DEBUG oslo_vmware.api [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350262, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.643371] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1193.643371] env[63355]: value = "task-1350263" [ 1193.643371] env[63355]: _type = "Task" [ 1193.643371] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.650369] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350263, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.867317] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.867501] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.867662] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1193.867768] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Rebuilding the list of instances to heal {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1194.145134] env[63355]: DEBUG oslo_vmware.api [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350262, 'name': ReconfigVM_Task, 'duration_secs': 0.13942} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.148040] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287772', 'volume_id': 'b65b95db-ab9c-4ae3-b295-22a877310891', 'name': 'volume-b65b95db-ab9c-4ae3-b295-22a877310891', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d332e8f-4c89-4050-b5b7-3cd034f9edfe', 'attached_at': '', 'detached_at': '', 'volume_id': 'b65b95db-ab9c-4ae3-b295-22a877310891', 'serial': 'b65b95db-ab9c-4ae3-b295-22a877310891'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1194.154634] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350263, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.408330] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1194.408492] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquired lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.408645] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Forcefully refreshing network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1194.408858] env[63355]: DEBUG nova.objects.instance [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lazy-loading 'info_cache' on Instance uuid 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.655498] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350263, 'name': PowerOnVM_Task, 'duration_secs': 0.98741} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.655782] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1194.655993] env[63355]: DEBUG nova.compute.manager [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1194.656750] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f5d5b8-db41-4775-ba9a-46515b797f84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.167194] env[63355]: INFO nova.compute.manager [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] bringing vm to original state: 'stopped' [ 1195.183852] env[63355]: DEBUG nova.objects.instance [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'flavor' on Instance uuid 8d332e8f-4c89-4050-b5b7-3cd034f9edfe {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.523277] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.691596] env[63355]: DEBUG oslo_concurrency.lockutils [None req-492b51b8-a243-435a-9d2d-f892f4fdf48b tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.746s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.692993] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.170s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.138375] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Updating instance_info_cache with network_info: [{"id": "2ea6bc23-1560-4552-be29-22a2d0d863ab", "address": "fa:16:3e:f8:7c:fc", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ea6bc23-15", "ovs_interfaceid": "2ea6bc23-1560-4552-be29-22a2d0d863ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.174859] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "1c054e10-f650-4b47-a233-3bf17482a9b8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.175600] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.175814] env[63355]: DEBUG nova.compute.manager [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1196.176759] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eea3f8e-bd4b-4507-8aa8-cf12f5c4316f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.183679] env[63355]: DEBUG nova.compute.manager [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63355) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1196.186768] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1196.186999] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c12e7314-dda8-49a5-99dd-5e9f7790dfe4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.193700] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1196.193700] env[63355]: value = "task-1350264" [ 1196.193700] env[63355]: _type = "Task" [ 1196.193700] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.197031] env[63355]: INFO nova.compute.manager [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Detaching volume 48dd3241-16cb-4df7-9a85-da673593e136 [ 1196.207883] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350264, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.233130] env[63355]: INFO nova.virt.block_device [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Attempting to driver detach volume 48dd3241-16cb-4df7-9a85-da673593e136 from mountpoint /dev/sdb [ 1196.233377] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1196.233567] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287769', 'volume_id': '48dd3241-16cb-4df7-9a85-da673593e136', 'name': 'volume-48dd3241-16cb-4df7-9a85-da673593e136', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d332e8f-4c89-4050-b5b7-3cd034f9edfe', 'attached_at': '', 'detached_at': '', 'volume_id': '48dd3241-16cb-4df7-9a85-da673593e136', 'serial': '48dd3241-16cb-4df7-9a85-da673593e136'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1196.234469] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443da858-268a-4bb1-8c58-83ff85883bdc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.258931] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82943c2-bc66-4168-9cf4-3d288c8f8169 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.266423] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96dfa49b-cf3d-486f-9a3c-cb72252608a7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.290949] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6aad79-6997-4da2-bcbf-4762671232e0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.306589] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] The volume has not been displaced from its original location: [datastore1] volume-48dd3241-16cb-4df7-9a85-da673593e136/volume-48dd3241-16cb-4df7-9a85-da673593e136.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1196.311989] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1196.312338] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8a5b17c-95a4-4f2e-a5ef-c3d95cdc1a16 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.330705] env[63355]: DEBUG oslo_vmware.api [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1196.330705] env[63355]: value = "task-1350265" [ 1196.330705] env[63355]: _type = "Task" [ 1196.330705] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.339031] env[63355]: DEBUG oslo_vmware.api [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350265, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.640776] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Releasing lock "refresh_cache-7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1196.640987] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Updated the network info_cache for instance {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1196.641238] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.641398] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.641544] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.641690] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.641829] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.641968] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._sync_power_states {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.703456] env[63355]: DEBUG oslo_vmware.api [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350264, 'name': PowerOffVM_Task, 'duration_secs': 0.26196} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.703735] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1196.703927] env[63355]: DEBUG nova.compute.manager [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1196.704699] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3b6727-c7db-47bc-bd67-3ee17997afda {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.841043] env[63355]: DEBUG oslo_vmware.api [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350265, 'name': ReconfigVM_Task, 'duration_secs': 0.289274} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.841361] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1196.845895] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d73e6f7-d17f-44fa-b61a-f091763a606e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.860893] env[63355]: DEBUG oslo_vmware.api [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1196.860893] env[63355]: value = "task-1350266" [ 1196.860893] env[63355]: _type = "Task" [ 1196.860893] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.868551] env[63355]: DEBUG oslo_vmware.api [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350266, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.147051] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Getting list of instances from cluster (obj){ [ 1197.147051] env[63355]: value = "domain-c8" [ 1197.147051] env[63355]: _type = "ClusterComputeResource" [ 1197.147051] env[63355]: } {{(pid=63355) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1197.147431] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829e3f3e-c3f2-450c-a201-b31f2c22515a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.163064] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Got total of 6 instances {{(pid=63355) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1197.163226] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1197.163418] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid 8db23054-4556-496a-a4b1-a24c71c30f50 {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1197.163575] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid 8d332e8f-4c89-4050-b5b7-3cd034f9edfe {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1197.163729] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid ddaab84c-5f1d-4122-9408-2043f30a2227 {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1197.163881] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid 1c054e10-f650-4b47-a233-3bf17482a9b8 {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1197.164474] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Triggering sync for uuid 500cef8e-f058-4597-8410-ff83bfab06eb {{(pid=63355) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1197.164474] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.164642] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.164855] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "8db23054-4556-496a-a4b1-a24c71c30f50" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.165054] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "8db23054-4556-496a-a4b1-a24c71c30f50" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.165304] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.165502] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "ddaab84c-5f1d-4122-9408-2043f30a2227" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.165681] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "ddaab84c-5f1d-4122-9408-2043f30a2227" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.165906] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "1c054e10-f650-4b47-a233-3bf17482a9b8" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.166123] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "500cef8e-f058-4597-8410-ff83bfab06eb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.166301] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "500cef8e-f058-4597-8410-ff83bfab06eb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.166493] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.166626] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1197.167308] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e2a8eb-e964-4ed7-8dc9-8b65b250ec64 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.171406] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f233dfc9-647d-40b5-89a7-57882b487375 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.173921] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a01448-4ab3-44e8-a99d-1d63803cdb33 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.176592] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c51b3ec-47de-4172-83e5-cbb71e2dc56f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.217198] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.042s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.218231] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.052s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.218231] env[63355]: INFO nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] During sync_power_state the instance has a pending task (powering-off). Skip. [ 1197.218434] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.370517] env[63355]: DEBUG oslo_vmware.api [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350266, 'name': ReconfigVM_Task, 'duration_secs': 0.145738} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.370829] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287769', 'volume_id': '48dd3241-16cb-4df7-9a85-da673593e136', 'name': 'volume-48dd3241-16cb-4df7-9a85-da673593e136', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d332e8f-4c89-4050-b5b7-3cd034f9edfe', 'attached_at': '', 'detached_at': '', 'volume_id': '48dd3241-16cb-4df7-9a85-da673593e136', 'serial': '48dd3241-16cb-4df7-9a85-da673593e136'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1197.685134] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.520s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.689792] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "8db23054-4556-496a-a4b1-a24c71c30f50" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.525s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.691263] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "500cef8e-f058-4597-8410-ff83bfab06eb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.525s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.692681] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "ddaab84c-5f1d-4122-9408-2043f30a2227" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.527s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.724160] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.724399] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.724580] env[63355]: DEBUG nova.objects.instance [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63355) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1197.911517] env[63355]: DEBUG nova.objects.instance [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'flavor' on Instance uuid 8d332e8f-4c89-4050-b5b7-3cd034f9edfe {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1198.092289] env[63355]: DEBUG oslo_concurrency.lockutils [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "1c054e10-f650-4b47-a233-3bf17482a9b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.092574] env[63355]: DEBUG oslo_concurrency.lockutils [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.092796] env[63355]: DEBUG oslo_concurrency.lockutils [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "1c054e10-f650-4b47-a233-3bf17482a9b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.092981] env[63355]: DEBUG oslo_concurrency.lockutils [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.093174] env[63355]: DEBUG oslo_concurrency.lockutils [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.095645] env[63355]: INFO nova.compute.manager [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Terminating instance [ 1198.097437] env[63355]: DEBUG nova.compute.manager [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1198.097671] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1198.098492] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f955f6ca-457c-423b-8b71-78f5cca5cc80 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.105945] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1198.106191] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-907c8630-f44e-4cf6-8ed5-8553dedc7a27 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.170714] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1198.170951] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1198.171168] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleting the datastore file [datastore2] 1c054e10-f650-4b47-a233-3bf17482a9b8 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1198.171430] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-436215ec-ae4f-4df9-8f0f-8fd6fa2468c4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.177811] env[63355]: DEBUG oslo_vmware.api [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1198.177811] env[63355]: value = "task-1350268" [ 1198.177811] env[63355]: _type = "Task" [ 1198.177811] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.185918] env[63355]: DEBUG oslo_vmware.api [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350268, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.689613] env[63355]: DEBUG oslo_vmware.api [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350268, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12354} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.690028] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1198.690336] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1198.690633] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1198.690929] env[63355]: INFO nova.compute.manager [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1198.691327] env[63355]: DEBUG oslo.service.loopingcall [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.691626] env[63355]: DEBUG nova.compute.manager [-] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1198.691773] env[63355]: DEBUG nova.network.neutron [-] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1198.732884] env[63355]: DEBUG oslo_concurrency.lockutils [None req-4b6ad29d-22f4-44d2-a61f-bb0a6ba16040 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.769980] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.919351] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f5e6dbfe-139c-408f-8290-384d451f0a45 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.226s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.920528] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.755s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.921478] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a879f1-06fb-41fd-9515-6865914e0f43 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.981942] env[63355]: DEBUG nova.compute.manager [req-e3f35b5d-e97f-40ab-8229-ee6aae3789a6 req-5be9ed39-596d-43cf-98a4-e7c8f681b93b service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Received event network-vif-deleted-417738e8-effd-45de-9c4b-b0eda6d5ccfc {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1198.982178] env[63355]: INFO nova.compute.manager [req-e3f35b5d-e97f-40ab-8229-ee6aae3789a6 req-5be9ed39-596d-43cf-98a4-e7c8f681b93b service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Neutron deleted interface 417738e8-effd-45de-9c4b-b0eda6d5ccfc; detaching it from the instance and deleting it from the info cache [ 1198.982360] env[63355]: DEBUG nova.network.neutron [req-e3f35b5d-e97f-40ab-8229-ee6aae3789a6 req-5be9ed39-596d-43cf-98a4-e7c8f681b93b service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.431386] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.511s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.431674] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.662s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.434489] env[63355]: DEBUG nova.network.neutron [-] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.484355] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-69d0ef94-a131-4638-bea4-9105e9882390 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.494156] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212e1a56-a118-4fa0-b364-105b9c4299e2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.520106] env[63355]: DEBUG nova.compute.manager [req-e3f35b5d-e97f-40ab-8229-ee6aae3789a6 req-5be9ed39-596d-43cf-98a4-e7c8f681b93b service nova] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Detach interface failed, port_id=417738e8-effd-45de-9c4b-b0eda6d5ccfc, reason: Instance 1c054e10-f650-4b47-a233-3bf17482a9b8 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1199.934787] env[63355]: INFO nova.compute.manager [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Detaching volume b65b95db-ab9c-4ae3-b295-22a877310891 [ 1199.936960] env[63355]: INFO nova.compute.manager [-] [instance: 1c054e10-f650-4b47-a233-3bf17482a9b8] Took 1.25 seconds to deallocate network for instance. [ 1199.969126] env[63355]: INFO nova.virt.block_device [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Attempting to driver detach volume b65b95db-ab9c-4ae3-b295-22a877310891 from mountpoint /dev/sdc [ 1199.969370] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1199.969561] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287772', 'volume_id': 'b65b95db-ab9c-4ae3-b295-22a877310891', 'name': 'volume-b65b95db-ab9c-4ae3-b295-22a877310891', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d332e8f-4c89-4050-b5b7-3cd034f9edfe', 'attached_at': '', 'detached_at': '', 'volume_id': 'b65b95db-ab9c-4ae3-b295-22a877310891', 'serial': 'b65b95db-ab9c-4ae3-b295-22a877310891'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1199.970496] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d624575b-2dde-42f3-b76d-58aa74c78764 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.992701] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae6797e-4ff8-42a9-89de-64e300868a69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.999039] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14afd84-8957-4a25-8a7c-1d7beedfe7c3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.019762] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66fd1fcf-d3d5-4413-a0a7-4bfd4adccd4a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.033368] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] The volume has not been displaced from its original location: [datastore2] volume-b65b95db-ab9c-4ae3-b295-22a877310891/volume-b65b95db-ab9c-4ae3-b295-22a877310891.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1200.038502] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Reconfiguring VM instance instance-00000068 to detach disk 2002 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1200.038991] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82763eae-f48b-4e78-884b-bcde613eecc2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.055924] env[63355]: DEBUG oslo_vmware.api [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1200.055924] env[63355]: value = "task-1350269" [ 1200.055924] env[63355]: _type = "Task" [ 1200.055924] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.063207] env[63355]: DEBUG oslo_vmware.api [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350269, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.445218] env[63355]: DEBUG oslo_concurrency.lockutils [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.445555] env[63355]: DEBUG oslo_concurrency.lockutils [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.445722] env[63355]: DEBUG nova.objects.instance [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lazy-loading 'resources' on Instance uuid 1c054e10-f650-4b47-a233-3bf17482a9b8 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1200.564965] env[63355]: DEBUG oslo_vmware.api [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350269, 'name': ReconfigVM_Task, 'duration_secs': 0.20694} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.565260] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Reconfigured VM instance instance-00000068 to detach disk 2002 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1200.569784] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ffc28ac9-9bc7-47e2-9468-b7a90c122a87 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.583662] env[63355]: DEBUG oslo_vmware.api [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1200.583662] env[63355]: value = "task-1350270" [ 1200.583662] env[63355]: _type = "Task" [ 1200.583662] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.590841] env[63355]: DEBUG oslo_vmware.api [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350270, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.027126] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8384527f-37fa-4c8a-9bb3-99b35d3873ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.034965] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a58c345-f9c1-4132-b706-11fc0d322024 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.064100] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5f666a-6723-4ea4-9625-5a535cba3980 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.070735] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ef044f-e0c8-41f9-af54-538523edf2eb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.084706] env[63355]: DEBUG nova.compute.provider_tree [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1201.092707] env[63355]: DEBUG oslo_vmware.api [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350270, 'name': ReconfigVM_Task, 'duration_secs': 0.126231} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.093504] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287772', 'volume_id': 'b65b95db-ab9c-4ae3-b295-22a877310891', 'name': 'volume-b65b95db-ab9c-4ae3-b295-22a877310891', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d332e8f-4c89-4050-b5b7-3cd034f9edfe', 'attached_at': '', 'detached_at': '', 'volume_id': 'b65b95db-ab9c-4ae3-b295-22a877310891', 'serial': 'b65b95db-ab9c-4ae3-b295-22a877310891'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1201.590411] env[63355]: DEBUG nova.scheduler.client.report [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1201.635758] env[63355]: DEBUG nova.objects.instance [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'flavor' on Instance uuid 8d332e8f-4c89-4050-b5b7-3cd034f9edfe {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1202.095799] env[63355]: DEBUG oslo_concurrency.lockutils [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.650s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.117308] env[63355]: INFO nova.scheduler.client.report [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted allocations for instance 1c054e10-f650-4b47-a233-3bf17482a9b8 [ 1202.624401] env[63355]: DEBUG oslo_concurrency.lockutils [None req-68df938e-9d33-4d73-8b73-09682605637e tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "1c054e10-f650-4b47-a233-3bf17482a9b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.532s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.643531] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f4f65a8d-3c7d-4112-af57-bb93c22765b9 tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.212s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.131041] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.131361] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.131580] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.131767] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.131995] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.134169] env[63355]: INFO nova.compute.manager [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Terminating instance [ 1203.137052] env[63355]: DEBUG nova.compute.manager [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1203.137052] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1203.137181] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa15f07-784f-4187-a583-709602d1602b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.145872] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1203.146164] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a54001e1-320c-49ca-a241-843d5799711e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.153630] env[63355]: DEBUG oslo_vmware.api [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1203.153630] env[63355]: value = "task-1350271" [ 1203.153630] env[63355]: _type = "Task" [ 1203.153630] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.168943] env[63355]: DEBUG oslo_vmware.api [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.663784] env[63355]: DEBUG oslo_vmware.api [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350271, 'name': PowerOffVM_Task, 'duration_secs': 0.182742} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.664098] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1203.664278] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1203.664533] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5032fee-14a0-4ae5-b502-bcf4ed1a6977 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.729971] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1203.730265] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1203.730458] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Deleting the datastore file [datastore2] 8d332e8f-4c89-4050-b5b7-3cd034f9edfe {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1203.730729] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-871d9ff4-e8de-4fd8-bb8b-715ead11fcf9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.738075] env[63355]: DEBUG oslo_vmware.api [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for the task: (returnval){ [ 1203.738075] env[63355]: value = "task-1350273" [ 1203.738075] env[63355]: _type = "Task" [ 1203.738075] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.745305] env[63355]: DEBUG oslo_vmware.api [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350273, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.248122] env[63355]: DEBUG oslo_vmware.api [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Task: {'id': task-1350273, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128697} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.248391] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1204.248579] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1204.248763] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1204.248968] env[63355]: INFO nova.compute.manager [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1204.249240] env[63355]: DEBUG oslo.service.loopingcall [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1204.249439] env[63355]: DEBUG nova.compute.manager [-] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1204.249534] env[63355]: DEBUG nova.network.neutron [-] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1204.722977] env[63355]: DEBUG nova.compute.manager [req-c478cd1b-efb7-45bd-894f-d7e1ab4257dd req-7546778c-1ca2-480e-b0cb-3e2c2262ae1b service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Received event network-vif-deleted-69950442-ccce-4a0a-95ad-cc51720ad0a6 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1204.723225] env[63355]: INFO nova.compute.manager [req-c478cd1b-efb7-45bd-894f-d7e1ab4257dd req-7546778c-1ca2-480e-b0cb-3e2c2262ae1b service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Neutron deleted interface 69950442-ccce-4a0a-95ad-cc51720ad0a6; detaching it from the instance and deleting it from the info cache [ 1204.723384] env[63355]: DEBUG nova.network.neutron [req-c478cd1b-efb7-45bd-894f-d7e1ab4257dd req-7546778c-1ca2-480e-b0cb-3e2c2262ae1b service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.037664] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "5b246e6a-9974-49e4-b213-7664c90e8420" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.037900] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.202879] env[63355]: DEBUG nova.network.neutron [-] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.225828] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-194438fc-c1af-4427-945e-eb6c328885c0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.235023] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1eb7511-3c81-4172-a2fb-4e99883aa60d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.259750] env[63355]: DEBUG nova.compute.manager [req-c478cd1b-efb7-45bd-894f-d7e1ab4257dd req-7546778c-1ca2-480e-b0cb-3e2c2262ae1b service nova] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Detach interface failed, port_id=69950442-ccce-4a0a-95ad-cc51720ad0a6, reason: Instance 8d332e8f-4c89-4050-b5b7-3cd034f9edfe could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1205.541986] env[63355]: DEBUG nova.compute.manager [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1205.705295] env[63355]: INFO nova.compute.manager [-] [instance: 8d332e8f-4c89-4050-b5b7-3cd034f9edfe] Took 1.46 seconds to deallocate network for instance. [ 1206.064367] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.064640] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.066562] env[63355]: INFO nova.compute.claims [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1206.211388] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.145857] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eff9c30-20ed-432a-a5e9-1744a24129ec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.153784] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b663a872-23b0-46a6-809d-81a38488f51b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.184125] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b12baeba-0bc8-4c10-ac24-453a8890f6a1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.191043] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5e0401-e419-4859-a44d-290c6237e651 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.203714] env[63355]: DEBUG nova.compute.provider_tree [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1207.707237] env[63355]: DEBUG nova.scheduler.client.report [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1208.212666] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.148s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.213241] env[63355]: DEBUG nova.compute.manager [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1208.215929] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.005s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.216536] env[63355]: DEBUG nova.objects.instance [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lazy-loading 'resources' on Instance uuid 8d332e8f-4c89-4050-b5b7-3cd034f9edfe {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.720062] env[63355]: DEBUG nova.compute.utils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1208.723770] env[63355]: DEBUG nova.compute.manager [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1208.723770] env[63355]: DEBUG nova.network.neutron [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1208.774038] env[63355]: DEBUG nova.policy [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a61bc098ea4e43ceab263ae9890da7c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff5b796e7bb0458dbd3b862dcae7bdd1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1208.798390] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e11d33-e91c-4fbc-a358-c2e5165e23a2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.807691] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee950c7-b196-4b7d-aab0-07f87a5c4b7a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.837107] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fab2cee-3150-44ae-83a0-b5db99e2c751 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.844173] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a928f319-dd5f-46e3-95fc-371a17c60d6b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.856741] env[63355]: DEBUG nova.compute.provider_tree [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.031662] env[63355]: DEBUG nova.network.neutron [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Successfully created port: 313f0c1d-9da1-41ec-bba9-5ff65d52db63 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1209.224946] env[63355]: DEBUG nova.compute.manager [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1209.359705] env[63355]: DEBUG nova.scheduler.client.report [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1209.864975] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.649s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.885658] env[63355]: INFO nova.scheduler.client.report [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Deleted allocations for instance 8d332e8f-4c89-4050-b5b7-3cd034f9edfe [ 1210.235255] env[63355]: DEBUG nova.compute.manager [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1210.260341] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1210.260624] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1210.260787] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1210.260973] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1210.261147] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1210.261302] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1210.261519] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1210.261682] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1210.261852] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1210.262608] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1210.262758] env[63355]: DEBUG nova.virt.hardware [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1210.263661] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f682da03-f351-41fa-9657-4514fa973152 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.272710] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0a147d-475f-47e8-b209-f144846224c8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.394836] env[63355]: DEBUG oslo_concurrency.lockutils [None req-cbbfa050-e2ff-4c1f-acd5-e24bc33353ac tempest-AttachVolumeTestJSON-1904511684 tempest-AttachVolumeTestJSON-1904511684-project-member] Lock "8d332e8f-4c89-4050-b5b7-3cd034f9edfe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.263s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.417151] env[63355]: DEBUG nova.compute.manager [req-d55a7cd2-db81-4c7e-ba9c-06b4499fe660 req-367472ac-47a6-427e-92c2-465367dc045b service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Received event network-vif-plugged-313f0c1d-9da1-41ec-bba9-5ff65d52db63 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1210.417394] env[63355]: DEBUG oslo_concurrency.lockutils [req-d55a7cd2-db81-4c7e-ba9c-06b4499fe660 req-367472ac-47a6-427e-92c2-465367dc045b service nova] Acquiring lock "5b246e6a-9974-49e4-b213-7664c90e8420-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.417619] env[63355]: DEBUG oslo_concurrency.lockutils [req-d55a7cd2-db81-4c7e-ba9c-06b4499fe660 req-367472ac-47a6-427e-92c2-465367dc045b service nova] Lock "5b246e6a-9974-49e4-b213-7664c90e8420-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.417791] env[63355]: DEBUG oslo_concurrency.lockutils [req-d55a7cd2-db81-4c7e-ba9c-06b4499fe660 req-367472ac-47a6-427e-92c2-465367dc045b service nova] Lock "5b246e6a-9974-49e4-b213-7664c90e8420-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.417986] env[63355]: DEBUG nova.compute.manager [req-d55a7cd2-db81-4c7e-ba9c-06b4499fe660 req-367472ac-47a6-427e-92c2-465367dc045b service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] No waiting events found dispatching network-vif-plugged-313f0c1d-9da1-41ec-bba9-5ff65d52db63 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1210.418131] env[63355]: WARNING nova.compute.manager [req-d55a7cd2-db81-4c7e-ba9c-06b4499fe660 req-367472ac-47a6-427e-92c2-465367dc045b service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Received unexpected event network-vif-plugged-313f0c1d-9da1-41ec-bba9-5ff65d52db63 for instance with vm_state building and task_state spawning. [ 1210.506185] env[63355]: DEBUG nova.network.neutron [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Successfully updated port: 313f0c1d-9da1-41ec-bba9-5ff65d52db63 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1211.009051] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "refresh_cache-5b246e6a-9974-49e4-b213-7664c90e8420" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1211.009051] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "refresh_cache-5b246e6a-9974-49e4-b213-7664c90e8420" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.009051] env[63355]: DEBUG nova.network.neutron [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1211.540292] env[63355]: DEBUG nova.network.neutron [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1211.662709] env[63355]: DEBUG nova.network.neutron [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Updating instance_info_cache with network_info: [{"id": "313f0c1d-9da1-41ec-bba9-5ff65d52db63", "address": "fa:16:3e:8f:e5:45", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap313f0c1d-9d", "ovs_interfaceid": "313f0c1d-9da1-41ec-bba9-5ff65d52db63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.164989] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "refresh_cache-5b246e6a-9974-49e4-b213-7664c90e8420" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1212.165367] env[63355]: DEBUG nova.compute.manager [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Instance network_info: |[{"id": "313f0c1d-9da1-41ec-bba9-5ff65d52db63", "address": "fa:16:3e:8f:e5:45", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap313f0c1d-9d", "ovs_interfaceid": "313f0c1d-9da1-41ec-bba9-5ff65d52db63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1212.165824] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:e5:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '313f0c1d-9da1-41ec-bba9-5ff65d52db63', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1212.173442] env[63355]: DEBUG oslo.service.loopingcall [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1212.173658] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1212.173915] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a30b167a-7878-43b5-b6f0-bbe7577f6ace {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.193238] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1212.193238] env[63355]: value = "task-1350275" [ 1212.193238] env[63355]: _type = "Task" [ 1212.193238] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.200325] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350275, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.444751] env[63355]: DEBUG nova.compute.manager [req-ee344c7d-5d81-4a08-887a-bc9c8c6c2179 req-3f6cd720-c234-4724-97ca-1f0cc289e363 service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Received event network-changed-313f0c1d-9da1-41ec-bba9-5ff65d52db63 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1212.444954] env[63355]: DEBUG nova.compute.manager [req-ee344c7d-5d81-4a08-887a-bc9c8c6c2179 req-3f6cd720-c234-4724-97ca-1f0cc289e363 service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Refreshing instance network info cache due to event network-changed-313f0c1d-9da1-41ec-bba9-5ff65d52db63. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1212.445203] env[63355]: DEBUG oslo_concurrency.lockutils [req-ee344c7d-5d81-4a08-887a-bc9c8c6c2179 req-3f6cd720-c234-4724-97ca-1f0cc289e363 service nova] Acquiring lock "refresh_cache-5b246e6a-9974-49e4-b213-7664c90e8420" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.445366] env[63355]: DEBUG oslo_concurrency.lockutils [req-ee344c7d-5d81-4a08-887a-bc9c8c6c2179 req-3f6cd720-c234-4724-97ca-1f0cc289e363 service nova] Acquired lock "refresh_cache-5b246e6a-9974-49e4-b213-7664c90e8420" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.445551] env[63355]: DEBUG nova.network.neutron [req-ee344c7d-5d81-4a08-887a-bc9c8c6c2179 req-3f6cd720-c234-4724-97ca-1f0cc289e363 service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Refreshing network info cache for port 313f0c1d-9da1-41ec-bba9-5ff65d52db63 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1212.702508] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350275, 'name': CreateVM_Task, 'duration_secs': 0.29771} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.702940] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1212.703754] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.703948] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.704288] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1212.704552] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31d7ca28-ee7b-40f8-afaa-f00b95b7b45f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.710078] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1212.710078] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525b10d2-1903-4766-9f6b-7bf1b2415307" [ 1212.710078] env[63355]: _type = "Task" [ 1212.710078] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.718839] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525b10d2-1903-4766-9f6b-7bf1b2415307, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.146479] env[63355]: DEBUG nova.network.neutron [req-ee344c7d-5d81-4a08-887a-bc9c8c6c2179 req-3f6cd720-c234-4724-97ca-1f0cc289e363 service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Updated VIF entry in instance network info cache for port 313f0c1d-9da1-41ec-bba9-5ff65d52db63. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1213.146867] env[63355]: DEBUG nova.network.neutron [req-ee344c7d-5d81-4a08-887a-bc9c8c6c2179 req-3f6cd720-c234-4724-97ca-1f0cc289e363 service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Updating instance_info_cache with network_info: [{"id": "313f0c1d-9da1-41ec-bba9-5ff65d52db63", "address": "fa:16:3e:8f:e5:45", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap313f0c1d-9d", "ovs_interfaceid": "313f0c1d-9da1-41ec-bba9-5ff65d52db63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.221118] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]525b10d2-1903-4766-9f6b-7bf1b2415307, 'name': SearchDatastore_Task, 'duration_secs': 0.00862} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.221405] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.221650] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1213.221923] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1213.222105] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.222401] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1213.222789] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-78325ed9-9d4e-4203-a472-fdfcd5e24deb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.230541] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1213.230713] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1213.231404] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b94f581-ca67-4738-9859-af716c38ad89 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.237239] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1213.237239] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523f7459-5c13-6e29-0de0-6627d4a08824" [ 1213.237239] env[63355]: _type = "Task" [ 1213.237239] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.244537] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523f7459-5c13-6e29-0de0-6627d4a08824, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.478358] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "ddaab84c-5f1d-4122-9408-2043f30a2227" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.478563] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "ddaab84c-5f1d-4122-9408-2043f30a2227" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.478779] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "ddaab84c-5f1d-4122-9408-2043f30a2227-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.478995] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "ddaab84c-5f1d-4122-9408-2043f30a2227-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.479181] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "ddaab84c-5f1d-4122-9408-2043f30a2227-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.481366] env[63355]: INFO nova.compute.manager [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Terminating instance [ 1213.483186] env[63355]: DEBUG nova.compute.manager [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1213.483379] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1213.484195] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7fce89-fafa-4bcf-8d71-7bb6addc0b42 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.491624] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1213.491880] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce5129b3-0a8a-4d4c-ae17-b935b0ad36c5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.498663] env[63355]: DEBUG oslo_vmware.api [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1213.498663] env[63355]: value = "task-1350277" [ 1213.498663] env[63355]: _type = "Task" [ 1213.498663] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.506253] env[63355]: DEBUG oslo_vmware.api [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.650184] env[63355]: DEBUG oslo_concurrency.lockutils [req-ee344c7d-5d81-4a08-887a-bc9c8c6c2179 req-3f6cd720-c234-4724-97ca-1f0cc289e363 service nova] Releasing lock "refresh_cache-5b246e6a-9974-49e4-b213-7664c90e8420" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.747373] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]523f7459-5c13-6e29-0de0-6627d4a08824, 'name': SearchDatastore_Task, 'duration_secs': 0.008206} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.748226] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de63a6af-1930-4abf-aa5b-4f72c80399df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.753760] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1213.753760] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d714a5-1e07-9665-4c3f-de64c4e4f9f5" [ 1213.753760] env[63355]: _type = "Task" [ 1213.753760] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.761700] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d714a5-1e07-9665-4c3f-de64c4e4f9f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.008138] env[63355]: DEBUG oslo_vmware.api [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350277, 'name': PowerOffVM_Task, 'duration_secs': 0.163007} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.008391] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1214.008538] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1214.008785] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df4ddcea-e2cf-49ab-a4e7-1829ec99c7ae {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.066423] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1214.066602] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1214.066796] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleting the datastore file [datastore1] ddaab84c-5f1d-4122-9408-2043f30a2227 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1214.067119] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87c2da9f-c830-411c-ba23-8633be23bfec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.073089] env[63355]: DEBUG oslo_vmware.api [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1214.073089] env[63355]: value = "task-1350279" [ 1214.073089] env[63355]: _type = "Task" [ 1214.073089] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.080151] env[63355]: DEBUG oslo_vmware.api [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350279, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.265135] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d714a5-1e07-9665-4c3f-de64c4e4f9f5, 'name': SearchDatastore_Task, 'duration_secs': 0.008748} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.265135] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1214.265135] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 5b246e6a-9974-49e4-b213-7664c90e8420/5b246e6a-9974-49e4-b213-7664c90e8420.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1214.265295] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34512b02-2037-461c-9d1a-f1ba333f973f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.272094] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1214.272094] env[63355]: value = "task-1350280" [ 1214.272094] env[63355]: _type = "Task" [ 1214.272094] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.279680] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350280, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.582694] env[63355]: DEBUG oslo_vmware.api [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350279, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130521} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.582694] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1214.582889] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1214.582993] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1214.583316] env[63355]: INFO nova.compute.manager [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1214.583617] env[63355]: DEBUG oslo.service.loopingcall [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1214.583862] env[63355]: DEBUG nova.compute.manager [-] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1214.583982] env[63355]: DEBUG nova.network.neutron [-] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1214.782038] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350280, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46261} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.782441] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 5b246e6a-9974-49e4-b213-7664c90e8420/5b246e6a-9974-49e4-b213-7664c90e8420.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1214.782547] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1214.782925] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c2a47b4-970f-433c-8420-345c595d4e16 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.789627] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1214.789627] env[63355]: value = "task-1350281" [ 1214.789627] env[63355]: _type = "Task" [ 1214.789627] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.797171] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350281, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.040294] env[63355]: DEBUG nova.compute.manager [req-6bf1870d-b4cc-4077-9083-47f07aa242bb req-7fd52be3-3498-4f17-9334-ed71f50d53f9 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Received event network-vif-deleted-bba4c0ca-7f31-49ce-92ec-703f913914ce {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1215.040375] env[63355]: INFO nova.compute.manager [req-6bf1870d-b4cc-4077-9083-47f07aa242bb req-7fd52be3-3498-4f17-9334-ed71f50d53f9 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Neutron deleted interface bba4c0ca-7f31-49ce-92ec-703f913914ce; detaching it from the instance and deleting it from the info cache [ 1215.040555] env[63355]: DEBUG nova.network.neutron [req-6bf1870d-b4cc-4077-9083-47f07aa242bb req-7fd52be3-3498-4f17-9334-ed71f50d53f9 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1215.299246] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350281, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111208} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.299637] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1215.300576] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa2a042-6428-48b6-aeec-2b1c466bcb13 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.323132] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 5b246e6a-9974-49e4-b213-7664c90e8420/5b246e6a-9974-49e4-b213-7664c90e8420.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1215.324630] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85ee3992-750f-4344-8981-e775f7cd1fc1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.343720] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1215.343720] env[63355]: value = "task-1350282" [ 1215.343720] env[63355]: _type = "Task" [ 1215.343720] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.352090] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350282, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.517873] env[63355]: DEBUG nova.network.neutron [-] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1215.543394] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7732b56-61af-4f76-b072-a41d51ec1216 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.553187] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f058fb6-9f30-462e-8ee9-03ff82b7e64b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.578637] env[63355]: DEBUG nova.compute.manager [req-6bf1870d-b4cc-4077-9083-47f07aa242bb req-7fd52be3-3498-4f17-9334-ed71f50d53f9 service nova] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Detach interface failed, port_id=bba4c0ca-7f31-49ce-92ec-703f913914ce, reason: Instance ddaab84c-5f1d-4122-9408-2043f30a2227 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1215.854482] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350282, 'name': ReconfigVM_Task, 'duration_secs': 0.26345} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.854790] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 5b246e6a-9974-49e4-b213-7664c90e8420/5b246e6a-9974-49e4-b213-7664c90e8420.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1215.855436] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8cd7a091-c125-4a36-b39f-c06bd18e1f50 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.862706] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1215.862706] env[63355]: value = "task-1350283" [ 1215.862706] env[63355]: _type = "Task" [ 1215.862706] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.870726] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350283, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.020726] env[63355]: INFO nova.compute.manager [-] [instance: ddaab84c-5f1d-4122-9408-2043f30a2227] Took 1.44 seconds to deallocate network for instance. [ 1216.372520] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350283, 'name': Rename_Task, 'duration_secs': 0.333654} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.372990] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1216.373347] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-426cf58a-c9a3-4890-bf1c-dc135d6aade3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.379367] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1216.379367] env[63355]: value = "task-1350285" [ 1216.379367] env[63355]: _type = "Task" [ 1216.379367] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.392521] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350285, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.527079] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.527434] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.527690] env[63355]: DEBUG nova.objects.instance [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lazy-loading 'resources' on Instance uuid ddaab84c-5f1d-4122-9408-2043f30a2227 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.888933] env[63355]: DEBUG oslo_vmware.api [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350285, 'name': PowerOnVM_Task, 'duration_secs': 0.403621} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.889320] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1216.889443] env[63355]: INFO nova.compute.manager [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1216.889623] env[63355]: DEBUG nova.compute.manager [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1216.890395] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25849784-8945-4aeb-8f05-c87dc2edca5c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.096544] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94b99dd-00d1-4811-afbb-7d21bc60264f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.104241] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310f2dd5-fcf5-40d9-86ff-919858959437 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.133536] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87dd1f3d-6a86-4e04-ae55-0132a71a7a9e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.140566] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7678688a-956f-45c7-bf14-d0b0f3f975ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.153506] env[63355]: DEBUG nova.compute.provider_tree [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.406545] env[63355]: INFO nova.compute.manager [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Took 11.36 seconds to build instance. [ 1217.656222] env[63355]: DEBUG nova.scheduler.client.report [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1217.908221] env[63355]: DEBUG oslo_concurrency.lockutils [None req-375d7cdd-011d-45ce-97dc-0acc649dc77b tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.870s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.045356] env[63355]: DEBUG nova.compute.manager [req-9d07a88c-e396-41f9-9e1f-1eb94bfe406c req-c9bcb1ad-e610-4210-b0c1-554034e04912 service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Received event network-changed-313f0c1d-9da1-41ec-bba9-5ff65d52db63 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1218.045568] env[63355]: DEBUG nova.compute.manager [req-9d07a88c-e396-41f9-9e1f-1eb94bfe406c req-c9bcb1ad-e610-4210-b0c1-554034e04912 service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Refreshing instance network info cache due to event network-changed-313f0c1d-9da1-41ec-bba9-5ff65d52db63. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1218.045781] env[63355]: DEBUG oslo_concurrency.lockutils [req-9d07a88c-e396-41f9-9e1f-1eb94bfe406c req-c9bcb1ad-e610-4210-b0c1-554034e04912 service nova] Acquiring lock "refresh_cache-5b246e6a-9974-49e4-b213-7664c90e8420" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.045928] env[63355]: DEBUG oslo_concurrency.lockutils [req-9d07a88c-e396-41f9-9e1f-1eb94bfe406c req-c9bcb1ad-e610-4210-b0c1-554034e04912 service nova] Acquired lock "refresh_cache-5b246e6a-9974-49e4-b213-7664c90e8420" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.046104] env[63355]: DEBUG nova.network.neutron [req-9d07a88c-e396-41f9-9e1f-1eb94bfe406c req-c9bcb1ad-e610-4210-b0c1-554034e04912 service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Refreshing network info cache for port 313f0c1d-9da1-41ec-bba9-5ff65d52db63 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1218.160590] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.179895] env[63355]: INFO nova.scheduler.client.report [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleted allocations for instance ddaab84c-5f1d-4122-9408-2043f30a2227 [ 1218.687720] env[63355]: DEBUG oslo_concurrency.lockutils [None req-c9916568-2121-404d-a87e-6f206d6788c3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "ddaab84c-5f1d-4122-9408-2043f30a2227" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.208s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.761616] env[63355]: DEBUG nova.network.neutron [req-9d07a88c-e396-41f9-9e1f-1eb94bfe406c req-c9bcb1ad-e610-4210-b0c1-554034e04912 service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Updated VIF entry in instance network info cache for port 313f0c1d-9da1-41ec-bba9-5ff65d52db63. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1218.761844] env[63355]: DEBUG nova.network.neutron [req-9d07a88c-e396-41f9-9e1f-1eb94bfe406c req-c9bcb1ad-e610-4210-b0c1-554034e04912 service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Updating instance_info_cache with network_info: [{"id": "313f0c1d-9da1-41ec-bba9-5ff65d52db63", "address": "fa:16:3e:8f:e5:45", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap313f0c1d-9d", "ovs_interfaceid": "313f0c1d-9da1-41ec-bba9-5ff65d52db63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.266508] env[63355]: DEBUG oslo_concurrency.lockutils [req-9d07a88c-e396-41f9-9e1f-1eb94bfe406c req-c9bcb1ad-e610-4210-b0c1-554034e04912 service nova] Releasing lock "refresh_cache-5b246e6a-9974-49e4-b213-7664c90e8420" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.393457] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.393819] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1219.897939] env[63355]: INFO nova.compute.manager [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Detaching volume 359e6f56-7a86-4095-95e0-d874eaf30d00 [ 1219.931438] env[63355]: INFO nova.virt.block_device [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Attempting to driver detach volume 359e6f56-7a86-4095-95e0-d874eaf30d00 from mountpoint /dev/sdb [ 1219.931724] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1219.931907] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287767', 'volume_id': '359e6f56-7a86-4095-95e0-d874eaf30d00', 'name': 'volume-359e6f56-7a86-4095-95e0-d874eaf30d00', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7aee05be-b8b4-4767-b5d4-88aa4a21d5cd', 'attached_at': '', 'detached_at': '', 'volume_id': '359e6f56-7a86-4095-95e0-d874eaf30d00', 'serial': '359e6f56-7a86-4095-95e0-d874eaf30d00'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1219.932923] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a606c96a-53a3-487d-828c-cfef4bc1938d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.955568] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac73448-bdd7-4718-a737-1e281aeaa47c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.963315] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd0afe7-093c-4755-8b09-dc8f68409da5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.986820] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462816b6-88d7-4f4a-a27e-998baba38717 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.002473] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] The volume has not been displaced from its original location: [datastore1] volume-359e6f56-7a86-4095-95e0-d874eaf30d00/volume-359e6f56-7a86-4095-95e0-d874eaf30d00.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1220.007697] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Reconfiguring VM instance instance-00000065 to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1220.008040] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-372bc253-6253-430a-b686-46fffc7ffb7e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.026245] env[63355]: DEBUG oslo_vmware.api [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1220.026245] env[63355]: value = "task-1350286" [ 1220.026245] env[63355]: _type = "Task" [ 1220.026245] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.035617] env[63355]: DEBUG oslo_vmware.api [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350286, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.536076] env[63355]: DEBUG oslo_vmware.api [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350286, 'name': ReconfigVM_Task, 'duration_secs': 0.314702} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.536076] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Reconfigured VM instance instance-00000065 to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1220.540374] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf65a5c1-0565-4c35-8113-035e492e7922 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.555255] env[63355]: DEBUG oslo_vmware.api [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1220.555255] env[63355]: value = "task-1350287" [ 1220.555255] env[63355]: _type = "Task" [ 1220.555255] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.563231] env[63355]: DEBUG oslo_vmware.api [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350287, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.065542] env[63355]: DEBUG oslo_vmware.api [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350287, 'name': ReconfigVM_Task, 'duration_secs': 0.143406} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.065860] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287767', 'volume_id': '359e6f56-7a86-4095-95e0-d874eaf30d00', 'name': 'volume-359e6f56-7a86-4095-95e0-d874eaf30d00', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7aee05be-b8b4-4767-b5d4-88aa4a21d5cd', 'attached_at': '', 'detached_at': '', 'volume_id': '359e6f56-7a86-4095-95e0-d874eaf30d00', 'serial': '359e6f56-7a86-4095-95e0-d874eaf30d00'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1221.609180] env[63355]: DEBUG nova.objects.instance [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lazy-loading 'flavor' on Instance uuid 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.616279] env[63355]: DEBUG oslo_concurrency.lockutils [None req-3f8301f0-dcce-4fa6-ab89-274a2eb43d9a tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.222s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.646618] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.646993] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.647118] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.647309] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.647486] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.649694] env[63355]: INFO nova.compute.manager [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Terminating instance [ 1223.651531] env[63355]: DEBUG nova.compute.manager [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1223.651720] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1223.652552] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65d4b18-8b7f-4d2b-8f73-c440b1e4c251 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.660496] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1223.660983] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5231e1e3-9ea9-422c-8e91-c208164345c0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.666422] env[63355]: DEBUG oslo_vmware.api [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1223.666422] env[63355]: value = "task-1350288" [ 1223.666422] env[63355]: _type = "Task" [ 1223.666422] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.674021] env[63355]: DEBUG oslo_vmware.api [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350288, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.176649] env[63355]: DEBUG oslo_vmware.api [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350288, 'name': PowerOffVM_Task, 'duration_secs': 0.201006} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.177844] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1224.177844] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1224.177844] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e346f4e0-0659-4bf8-b327-2b860e5765dc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.242523] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1224.242771] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1224.242939] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleting the datastore file [datastore2] 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1224.243245] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de214d33-1f8b-49da-8f6d-eed7dc0da52c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.251135] env[63355]: DEBUG oslo_vmware.api [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1224.251135] env[63355]: value = "task-1350290" [ 1224.251135] env[63355]: _type = "Task" [ 1224.251135] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.258131] env[63355]: DEBUG oslo_vmware.api [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350290, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.760586] env[63355]: DEBUG oslo_vmware.api [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350290, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152074} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.760965] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1224.761031] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1224.761219] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1224.761403] env[63355]: INFO nova.compute.manager [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1224.761643] env[63355]: DEBUG oslo.service.loopingcall [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1224.761837] env[63355]: DEBUG nova.compute.manager [-] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1224.761924] env[63355]: DEBUG nova.network.neutron [-] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1225.213151] env[63355]: DEBUG nova.compute.manager [req-f54acaea-a09f-402e-93c5-5009fee27847 req-54519014-d2fb-4e35-9e6b-bedae207e8eb service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Received event network-vif-deleted-2ea6bc23-1560-4552-be29-22a2d0d863ab {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1225.213413] env[63355]: INFO nova.compute.manager [req-f54acaea-a09f-402e-93c5-5009fee27847 req-54519014-d2fb-4e35-9e6b-bedae207e8eb service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Neutron deleted interface 2ea6bc23-1560-4552-be29-22a2d0d863ab; detaching it from the instance and deleting it from the info cache [ 1225.213572] env[63355]: DEBUG nova.network.neutron [req-f54acaea-a09f-402e-93c5-5009fee27847 req-54519014-d2fb-4e35-9e6b-bedae207e8eb service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1225.685155] env[63355]: DEBUG nova.network.neutron [-] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1225.716154] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90c796b3-7427-480f-ac61-1152a5314f3b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.726120] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c336f5a2-439f-477f-a860-a45ea703308f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.754907] env[63355]: DEBUG nova.compute.manager [req-f54acaea-a09f-402e-93c5-5009fee27847 req-54519014-d2fb-4e35-9e6b-bedae207e8eb service nova] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Detach interface failed, port_id=2ea6bc23-1560-4552-be29-22a2d0d863ab, reason: Instance 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1226.188310] env[63355]: INFO nova.compute.manager [-] [instance: 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd] Took 1.43 seconds to deallocate network for instance. [ 1226.241528] env[63355]: DEBUG oslo_concurrency.lockutils [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "500cef8e-f058-4597-8410-ff83bfab06eb" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.241816] env[63355]: DEBUG oslo_concurrency.lockutils [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "500cef8e-f058-4597-8410-ff83bfab06eb" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.242012] env[63355]: INFO nova.compute.manager [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Shelving [ 1226.695323] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.695621] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.695895] env[63355]: DEBUG nova.objects.instance [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lazy-loading 'resources' on Instance uuid 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1226.749350] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1226.749561] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2926cea6-d1e6-45b3-b4de-6456af4f29dc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.757158] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1226.757158] env[63355]: value = "task-1350291" [ 1226.757158] env[63355]: _type = "Task" [ 1226.757158] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.765191] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350291, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.259035] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bea5721-acad-4192-8cb4-62c0e82571cb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.272177] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47369c88-49f1-4c5d-9d41-923a63f64156 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.275153] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350291, 'name': PowerOffVM_Task, 'duration_secs': 0.169864} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.275409] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1227.276395] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b4d1df-9d48-4b27-be71-070dbd47919b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.301685] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d61963d-da90-4479-a53e-24a615ab86a4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.316574] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3362fc-9880-43b0-aee7-fbe56e52226a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.321676] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c063e9-2e6c-4cc8-a5d2-a4e9744259c8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.337451] env[63355]: DEBUG nova.compute.provider_tree [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1227.831488] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Creating Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1227.831831] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5b1c2f56-4932-4657-93e4-d634ba7d8690 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.838773] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1227.838773] env[63355]: value = "task-1350292" [ 1227.838773] env[63355]: _type = "Task" [ 1227.838773] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.839664] env[63355]: DEBUG nova.scheduler.client.report [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1227.853411] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350292, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.348612] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350292, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.349356] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.654s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1228.374222] env[63355]: INFO nova.scheduler.client.report [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleted allocations for instance 7aee05be-b8b4-4767-b5d4-88aa4a21d5cd [ 1228.850070] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350292, 'name': CreateSnapshot_Task, 'duration_secs': 0.532808} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.850389] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Created Snapshot of the VM instance {{(pid=63355) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1228.851157] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b38b756-6379-431e-b088-691c7c0ec91a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.882853] env[63355]: DEBUG oslo_concurrency.lockutils [None req-12fd74af-c67f-4ebe-bd41-3e2e77604ce1 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "7aee05be-b8b4-4767-b5d4-88aa4a21d5cd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.236s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.369975] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Creating linked-clone VM from snapshot {{(pid=63355) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1229.369975] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5f8ace5c-ea2a-4705-8c52-f87bdd960e70 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.377521] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1229.377521] env[63355]: value = "task-1350293" [ 1229.377521] env[63355]: _type = "Task" [ 1229.377521] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.385769] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350293, 'name': CloneVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.887910] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350293, 'name': CloneVM_Task} progress is 94%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.389050] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350293, 'name': CloneVM_Task, 'duration_secs': 0.976463} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.389050] env[63355]: INFO nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Created linked-clone VM from snapshot [ 1230.389560] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca8794a-1d5c-4615-a679-5f0ffb5cd03a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.395996] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Uploading image f25b0300-4805-4fd8-9157-fefa86c07e1a {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1230.422250] env[63355]: DEBUG oslo_vmware.rw_handles [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1230.422250] env[63355]: value = "vm-287776" [ 1230.422250] env[63355]: _type = "VirtualMachine" [ 1230.422250] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1230.422590] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2acf6820-acd8-42d4-95f2-a90310aed1eb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.430136] env[63355]: DEBUG oslo_vmware.rw_handles [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lease: (returnval){ [ 1230.430136] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52086c3a-ca01-6d6c-df9c-e498ae0f331c" [ 1230.430136] env[63355]: _type = "HttpNfcLease" [ 1230.430136] env[63355]: } obtained for exporting VM: (result){ [ 1230.430136] env[63355]: value = "vm-287776" [ 1230.430136] env[63355]: _type = "VirtualMachine" [ 1230.430136] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1230.430479] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the lease: (returnval){ [ 1230.430479] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52086c3a-ca01-6d6c-df9c-e498ae0f331c" [ 1230.430479] env[63355]: _type = "HttpNfcLease" [ 1230.430479] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1230.437137] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1230.437137] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52086c3a-ca01-6d6c-df9c-e498ae0f331c" [ 1230.437137] env[63355]: _type = "HttpNfcLease" [ 1230.437137] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1230.743144] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "e4c425fa-0b13-456c-b12b-c1ff18593536" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.743389] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.938392] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1230.938392] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52086c3a-ca01-6d6c-df9c-e498ae0f331c" [ 1230.938392] env[63355]: _type = "HttpNfcLease" [ 1230.938392] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1230.938708] env[63355]: DEBUG oslo_vmware.rw_handles [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1230.938708] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52086c3a-ca01-6d6c-df9c-e498ae0f331c" [ 1230.938708] env[63355]: _type = "HttpNfcLease" [ 1230.938708] env[63355]: }. {{(pid=63355) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1230.939495] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf117fc4-68d2-4621-a00d-c966161ee633 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.946419] env[63355]: DEBUG oslo_vmware.rw_handles [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524cdce4-c787-10c4-5da4-e9fcc6a59b89/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1230.946628] env[63355]: DEBUG oslo_vmware.rw_handles [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524cdce4-c787-10c4-5da4-e9fcc6a59b89/disk-0.vmdk for reading. {{(pid=63355) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1231.036260] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b45c6739-5583-458d-9a44-677c6eb118f6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.246155] env[63355]: DEBUG nova.compute.manager [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1231.771619] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.771932] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.773759] env[63355]: INFO nova.compute.claims [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1232.844541] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60298f96-5cb4-4a88-bda6-e770e9a092c6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.853150] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1672d2-5170-4653-91f7-011ee73e2162 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.882809] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beef6a91-b431-4042-a562-c7e969d36740 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.890230] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20042734-7f29-4d81-bcc9-31849cea5e16 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.903573] env[63355]: DEBUG nova.compute.provider_tree [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.407217] env[63355]: DEBUG nova.scheduler.client.report [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1233.913585] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.141s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1233.914177] env[63355]: DEBUG nova.compute.manager [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1234.419433] env[63355]: DEBUG nova.compute.utils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1234.421089] env[63355]: DEBUG nova.compute.manager [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1234.421089] env[63355]: DEBUG nova.network.neutron [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1234.461383] env[63355]: DEBUG nova.policy [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7302e68a14984b08aef9c3af0d2a12ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '876995f179ed46b397822fa1be08ea29', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1234.742090] env[63355]: DEBUG nova.network.neutron [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Successfully created port: bc7a8f76-e220-4474-8089-83364388f489 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1234.924186] env[63355]: DEBUG nova.compute.manager [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1235.934422] env[63355]: DEBUG nova.compute.manager [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1235.959478] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1235.959748] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1235.959914] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1235.960114] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1235.960318] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1235.960506] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1235.960726] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1235.960889] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1235.961080] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1235.961285] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1235.961469] env[63355]: DEBUG nova.virt.hardware [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1235.962448] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cd947a-1dec-4b01-9e14-0311ea5fcf06 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.970725] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9288caf8-881a-48e0-a3d4-4c2ebd63c5e6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.127185] env[63355]: DEBUG nova.compute.manager [req-2cff78e2-143f-421c-996a-24bc708beebf req-99bc2848-af44-4b8a-b1fd-85c0597d3e23 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Received event network-vif-plugged-bc7a8f76-e220-4474-8089-83364388f489 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1236.127416] env[63355]: DEBUG oslo_concurrency.lockutils [req-2cff78e2-143f-421c-996a-24bc708beebf req-99bc2848-af44-4b8a-b1fd-85c0597d3e23 service nova] Acquiring lock "e4c425fa-0b13-456c-b12b-c1ff18593536-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1236.127625] env[63355]: DEBUG oslo_concurrency.lockutils [req-2cff78e2-143f-421c-996a-24bc708beebf req-99bc2848-af44-4b8a-b1fd-85c0597d3e23 service nova] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1236.127816] env[63355]: DEBUG oslo_concurrency.lockutils [req-2cff78e2-143f-421c-996a-24bc708beebf req-99bc2848-af44-4b8a-b1fd-85c0597d3e23 service nova] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.127952] env[63355]: DEBUG nova.compute.manager [req-2cff78e2-143f-421c-996a-24bc708beebf req-99bc2848-af44-4b8a-b1fd-85c0597d3e23 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] No waiting events found dispatching network-vif-plugged-bc7a8f76-e220-4474-8089-83364388f489 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1236.128134] env[63355]: WARNING nova.compute.manager [req-2cff78e2-143f-421c-996a-24bc708beebf req-99bc2848-af44-4b8a-b1fd-85c0597d3e23 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Received unexpected event network-vif-plugged-bc7a8f76-e220-4474-8089-83364388f489 for instance with vm_state building and task_state spawning. [ 1236.216315] env[63355]: DEBUG nova.network.neutron [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Successfully updated port: bc7a8f76-e220-4474-8089-83364388f489 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1236.720064] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "refresh_cache-e4c425fa-0b13-456c-b12b-c1ff18593536" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1236.720064] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "refresh_cache-e4c425fa-0b13-456c-b12b-c1ff18593536" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1236.720064] env[63355]: DEBUG nova.network.neutron [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1237.251471] env[63355]: DEBUG nova.network.neutron [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1237.379392] env[63355]: DEBUG nova.network.neutron [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Updating instance_info_cache with network_info: [{"id": "bc7a8f76-e220-4474-8089-83364388f489", "address": "fa:16:3e:0d:c1:e0", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc7a8f76-e2", "ovs_interfaceid": "bc7a8f76-e220-4474-8089-83364388f489", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.883842] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "refresh_cache-e4c425fa-0b13-456c-b12b-c1ff18593536" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1237.883842] env[63355]: DEBUG nova.compute.manager [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Instance network_info: |[{"id": "bc7a8f76-e220-4474-8089-83364388f489", "address": "fa:16:3e:0d:c1:e0", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc7a8f76-e2", "ovs_interfaceid": "bc7a8f76-e220-4474-8089-83364388f489", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1237.884119] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:c1:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc7a8f76-e220-4474-8089-83364388f489', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1237.892166] env[63355]: DEBUG oslo.service.loopingcall [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1237.892401] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1237.892999] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5aeb6a15-324e-45f8-b919-244e5bc1b701 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.915826] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1237.915826] env[63355]: value = "task-1350295" [ 1237.915826] env[63355]: _type = "Task" [ 1237.915826] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.926034] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350295, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.928550] env[63355]: DEBUG oslo_vmware.rw_handles [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524cdce4-c787-10c4-5da4-e9fcc6a59b89/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1237.929390] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01584ba-a8c6-438c-b158-6de8877dadf0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.934789] env[63355]: DEBUG oslo_vmware.rw_handles [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524cdce4-c787-10c4-5da4-e9fcc6a59b89/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1237.934956] env[63355]: ERROR oslo_vmware.rw_handles [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524cdce4-c787-10c4-5da4-e9fcc6a59b89/disk-0.vmdk due to incomplete transfer. [ 1237.935183] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-fa1aaed6-f762-4c28-af3b-124f67b60992 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.941865] env[63355]: DEBUG oslo_vmware.rw_handles [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524cdce4-c787-10c4-5da4-e9fcc6a59b89/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1237.942079] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Uploaded image f25b0300-4805-4fd8-9157-fefa86c07e1a to the Glance image server {{(pid=63355) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1237.944336] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Destroying the VM {{(pid=63355) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1237.944561] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1cff41d0-3778-4f5c-88e2-1fc3598bc623 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.949953] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1237.949953] env[63355]: value = "task-1350296" [ 1237.949953] env[63355]: _type = "Task" [ 1237.949953] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.957424] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350296, 'name': Destroy_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.153850] env[63355]: DEBUG nova.compute.manager [req-84ac1259-0f83-47d4-97aa-b54976014972 req-28e20867-b4c6-4c25-86ba-67115bccabe4 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Received event network-changed-bc7a8f76-e220-4474-8089-83364388f489 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1238.154153] env[63355]: DEBUG nova.compute.manager [req-84ac1259-0f83-47d4-97aa-b54976014972 req-28e20867-b4c6-4c25-86ba-67115bccabe4 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Refreshing instance network info cache due to event network-changed-bc7a8f76-e220-4474-8089-83364388f489. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1238.154447] env[63355]: DEBUG oslo_concurrency.lockutils [req-84ac1259-0f83-47d4-97aa-b54976014972 req-28e20867-b4c6-4c25-86ba-67115bccabe4 service nova] Acquiring lock "refresh_cache-e4c425fa-0b13-456c-b12b-c1ff18593536" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1238.154642] env[63355]: DEBUG oslo_concurrency.lockutils [req-84ac1259-0f83-47d4-97aa-b54976014972 req-28e20867-b4c6-4c25-86ba-67115bccabe4 service nova] Acquired lock "refresh_cache-e4c425fa-0b13-456c-b12b-c1ff18593536" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.154841] env[63355]: DEBUG nova.network.neutron [req-84ac1259-0f83-47d4-97aa-b54976014972 req-28e20867-b4c6-4c25-86ba-67115bccabe4 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Refreshing network info cache for port bc7a8f76-e220-4474-8089-83364388f489 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1238.425852] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350295, 'name': CreateVM_Task, 'duration_secs': 0.377143} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.426230] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1238.426701] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1238.426889] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.427240] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1238.427483] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f57453af-2b95-4349-8d4f-2253e7d93ae7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.431912] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1238.431912] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524a1b93-504b-e76b-8e0b-d0af10b76aa8" [ 1238.431912] env[63355]: _type = "Task" [ 1238.431912] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.438944] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524a1b93-504b-e76b-8e0b-d0af10b76aa8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.458299] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350296, 'name': Destroy_Task, 'duration_secs': 0.435438} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.458299] env[63355]: INFO nova.virt.vmwareapi.vm_util [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Destroyed the VM [ 1238.458532] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Deleting Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1238.458763] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-196708d3-dd2e-46ea-a777-e109540b458f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.464250] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1238.464250] env[63355]: value = "task-1350297" [ 1238.464250] env[63355]: _type = "Task" [ 1238.464250] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.471070] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350297, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.844200] env[63355]: DEBUG nova.network.neutron [req-84ac1259-0f83-47d4-97aa-b54976014972 req-28e20867-b4c6-4c25-86ba-67115bccabe4 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Updated VIF entry in instance network info cache for port bc7a8f76-e220-4474-8089-83364388f489. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1238.844566] env[63355]: DEBUG nova.network.neutron [req-84ac1259-0f83-47d4-97aa-b54976014972 req-28e20867-b4c6-4c25-86ba-67115bccabe4 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Updating instance_info_cache with network_info: [{"id": "bc7a8f76-e220-4474-8089-83364388f489", "address": "fa:16:3e:0d:c1:e0", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc7a8f76-e2", "ovs_interfaceid": "bc7a8f76-e220-4474-8089-83364388f489", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1238.942125] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]524a1b93-504b-e76b-8e0b-d0af10b76aa8, 'name': SearchDatastore_Task, 'duration_secs': 0.009002} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.942461] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1238.942671] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1238.942908] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1238.943088] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.943280] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1238.943546] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1fee3710-c304-4098-a058-2bebda808026 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.951697] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1238.951905] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1238.952625] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8863ce79-47ee-464b-8e37-86ae83197aef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.957462] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1238.957462] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529c1e82-89df-1482-75bc-4c6edba97226" [ 1238.957462] env[63355]: _type = "Task" [ 1238.957462] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.964443] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529c1e82-89df-1482-75bc-4c6edba97226, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.971077] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350297, 'name': RemoveSnapshot_Task, 'duration_secs': 0.324769} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.971317] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Deleted Snapshot of the VM instance {{(pid=63355) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1238.971608] env[63355]: DEBUG nova.compute.manager [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1238.972358] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c3c7de-bfbb-4296-bf3e-279e2cf2274a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.347927] env[63355]: DEBUG oslo_concurrency.lockutils [req-84ac1259-0f83-47d4-97aa-b54976014972 req-28e20867-b4c6-4c25-86ba-67115bccabe4 service nova] Releasing lock "refresh_cache-e4c425fa-0b13-456c-b12b-c1ff18593536" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1239.468057] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529c1e82-89df-1482-75bc-4c6edba97226, 'name': SearchDatastore_Task, 'duration_secs': 0.007768} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.468432] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bed767b-fc25-4b14-b26b-cef1de5df774 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.473679] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1239.473679] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d62110-810b-3c06-87c5-87177ac9349e" [ 1239.473679] env[63355]: _type = "Task" [ 1239.473679] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.482727] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d62110-810b-3c06-87c5-87177ac9349e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.483171] env[63355]: INFO nova.compute.manager [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Shelve offloading [ 1239.484789] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1239.485008] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b8244f1-87d3-431b-bf0c-6125e888e666 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.491509] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1239.491509] env[63355]: value = "task-1350298" [ 1239.491509] env[63355]: _type = "Task" [ 1239.491509] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.499346] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350298, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.983793] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52d62110-810b-3c06-87c5-87177ac9349e, 'name': SearchDatastore_Task, 'duration_secs': 0.008909} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.984037] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1239.984323] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] e4c425fa-0b13-456c-b12b-c1ff18593536/e4c425fa-0b13-456c-b12b-c1ff18593536.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1239.984576] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b21c8cc-11ea-4998-987f-badb421432fb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.990797] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1239.990797] env[63355]: value = "task-1350299" [ 1239.990797] env[63355]: _type = "Task" [ 1239.990797] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.000716] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350299, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.003893] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] VM already powered off {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1240.004089] env[63355]: DEBUG nova.compute.manager [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1240.004778] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4fe599-b203-4a53-b0e9-ef6d7fbd9c86 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.009801] env[63355]: DEBUG oslo_concurrency.lockutils [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1240.009965] env[63355]: DEBUG oslo_concurrency.lockutils [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.010149] env[63355]: DEBUG nova.network.neutron [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1240.500401] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350299, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.427327} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.500752] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] e4c425fa-0b13-456c-b12b-c1ff18593536/e4c425fa-0b13-456c-b12b-c1ff18593536.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1240.500896] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1240.501116] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e0e17d6e-1692-48d6-afbf-e13194f68704 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.506591] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1240.506591] env[63355]: value = "task-1350300" [ 1240.506591] env[63355]: _type = "Task" [ 1240.506591] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.515275] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350300, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.727720] env[63355]: DEBUG nova.network.neutron [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updating instance_info_cache with network_info: [{"id": "57106266-129a-4058-8690-6deccdf8ab52", "address": "fa:16:3e:9e:a7:c8", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57106266-12", "ovs_interfaceid": "57106266-129a-4058-8690-6deccdf8ab52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1241.016434] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350300, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061275} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.016768] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1241.017538] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed85aeaa-84de-4cd3-9d93-624354cc6eb4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.039490] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] e4c425fa-0b13-456c-b12b-c1ff18593536/e4c425fa-0b13-456c-b12b-c1ff18593536.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1241.039819] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a9877dd-cafc-460f-8e86-855ce77c25c7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.059666] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1241.059666] env[63355]: value = "task-1350301" [ 1241.059666] env[63355]: _type = "Task" [ 1241.059666] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.068190] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350301, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.231157] env[63355]: DEBUG oslo_concurrency.lockutils [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1241.486491] env[63355]: DEBUG nova.compute.manager [req-342799c3-9bcb-4a1c-917b-ee2278fbad54 req-86e0c4fc-cb91-4d54-8b94-deb3242f0dc3 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received event network-vif-unplugged-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1241.486719] env[63355]: DEBUG oslo_concurrency.lockutils [req-342799c3-9bcb-4a1c-917b-ee2278fbad54 req-86e0c4fc-cb91-4d54-8b94-deb3242f0dc3 service nova] Acquiring lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1241.486933] env[63355]: DEBUG oslo_concurrency.lockutils [req-342799c3-9bcb-4a1c-917b-ee2278fbad54 req-86e0c4fc-cb91-4d54-8b94-deb3242f0dc3 service nova] Lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1241.487119] env[63355]: DEBUG oslo_concurrency.lockutils [req-342799c3-9bcb-4a1c-917b-ee2278fbad54 req-86e0c4fc-cb91-4d54-8b94-deb3242f0dc3 service nova] Lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1241.487293] env[63355]: DEBUG nova.compute.manager [req-342799c3-9bcb-4a1c-917b-ee2278fbad54 req-86e0c4fc-cb91-4d54-8b94-deb3242f0dc3 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] No waiting events found dispatching network-vif-unplugged-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1241.487468] env[63355]: WARNING nova.compute.manager [req-342799c3-9bcb-4a1c-917b-ee2278fbad54 req-86e0c4fc-cb91-4d54-8b94-deb3242f0dc3 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received unexpected event network-vif-unplugged-57106266-129a-4058-8690-6deccdf8ab52 for instance with vm_state shelved and task_state shelving_offloading. [ 1241.565720] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1241.566638] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9d2d2a-fa68-41a3-955d-bc103af40a17 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.573096] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350301, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.577242] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1241.577514] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e34f21f7-c1c6-4cc5-9878-09c4cae66c9f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.647017] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1241.647304] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1241.647476] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleting the datastore file [datastore1] 500cef8e-f058-4597-8410-ff83bfab06eb {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1241.647761] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de960396-060c-4820-93fd-ba525a36f5ed {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.654266] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1241.654266] env[63355]: value = "task-1350303" [ 1241.654266] env[63355]: _type = "Task" [ 1241.654266] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.663865] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350303, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.070325] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350301, 'name': ReconfigVM_Task, 'duration_secs': 0.565762} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.070609] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Reconfigured VM instance instance-0000006d to attach disk [datastore2] e4c425fa-0b13-456c-b12b-c1ff18593536/e4c425fa-0b13-456c-b12b-c1ff18593536.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1242.071254] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0a3bd84b-c18f-4a09-b552-70669e7c0fd1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.077468] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1242.077468] env[63355]: value = "task-1350304" [ 1242.077468] env[63355]: _type = "Task" [ 1242.077468] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.085117] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350304, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.164300] env[63355]: DEBUG oslo_vmware.api [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350303, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150015} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.164626] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1242.164833] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1242.165045] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1242.190986] env[63355]: INFO nova.scheduler.client.report [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleted allocations for instance 500cef8e-f058-4597-8410-ff83bfab06eb [ 1242.588548] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350304, 'name': Rename_Task, 'duration_secs': 0.137301} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.588548] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1242.588548] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b209af9e-1832-45cc-9325-89ae791fbb09 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.593786] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1242.593786] env[63355]: value = "task-1350305" [ 1242.593786] env[63355]: _type = "Task" [ 1242.593786] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.601076] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350305, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.695596] env[63355]: DEBUG oslo_concurrency.lockutils [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1242.695862] env[63355]: DEBUG oslo_concurrency.lockutils [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1242.696164] env[63355]: DEBUG nova.objects.instance [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'resources' on Instance uuid 500cef8e-f058-4597-8410-ff83bfab06eb {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1243.105447] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350305, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.198479] env[63355]: DEBUG nova.objects.instance [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'numa_topology' on Instance uuid 500cef8e-f058-4597-8410-ff83bfab06eb {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1243.513302] env[63355]: DEBUG nova.compute.manager [req-6039dfcd-5bd5-4f3f-9f89-24d8e3af73de req-7df3334f-5ed8-4c0f-b183-7119fd801c4c service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received event network-changed-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1243.513492] env[63355]: DEBUG nova.compute.manager [req-6039dfcd-5bd5-4f3f-9f89-24d8e3af73de req-7df3334f-5ed8-4c0f-b183-7119fd801c4c service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Refreshing instance network info cache due to event network-changed-57106266-129a-4058-8690-6deccdf8ab52. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1243.513708] env[63355]: DEBUG oslo_concurrency.lockutils [req-6039dfcd-5bd5-4f3f-9f89-24d8e3af73de req-7df3334f-5ed8-4c0f-b183-7119fd801c4c service nova] Acquiring lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1243.513852] env[63355]: DEBUG oslo_concurrency.lockutils [req-6039dfcd-5bd5-4f3f-9f89-24d8e3af73de req-7df3334f-5ed8-4c0f-b183-7119fd801c4c service nova] Acquired lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.514021] env[63355]: DEBUG nova.network.neutron [req-6039dfcd-5bd5-4f3f-9f89-24d8e3af73de req-7df3334f-5ed8-4c0f-b183-7119fd801c4c service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Refreshing network info cache for port 57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1243.603917] env[63355]: DEBUG oslo_vmware.api [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350305, 'name': PowerOnVM_Task, 'duration_secs': 0.547878} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.604237] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1243.604383] env[63355]: INFO nova.compute.manager [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Took 7.67 seconds to spawn the instance on the hypervisor. [ 1243.604595] env[63355]: DEBUG nova.compute.manager [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1243.605363] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8fd903-7c60-4fe6-a52d-7f10efad17da {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.700543] env[63355]: DEBUG nova.objects.base [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Object Instance<500cef8e-f058-4597-8410-ff83bfab06eb> lazy-loaded attributes: resources,numa_topology {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1243.754422] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b519b6-a555-49a8-98c3-44b173e4db97 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.761941] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6dc6277-5628-4d86-b519-f74ab7b1200e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.790942] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a692726-eb0f-4552-8973-f47ee0356623 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.797682] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b53862-8a86-4971-95ab-b46f7f571a43 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.810571] env[63355]: DEBUG nova.compute.provider_tree [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1243.841168] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "500cef8e-f058-4597-8410-ff83bfab06eb" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.122863] env[63355]: INFO nova.compute.manager [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Took 12.37 seconds to build instance. [ 1244.230247] env[63355]: DEBUG nova.network.neutron [req-6039dfcd-5bd5-4f3f-9f89-24d8e3af73de req-7df3334f-5ed8-4c0f-b183-7119fd801c4c service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updated VIF entry in instance network info cache for port 57106266-129a-4058-8690-6deccdf8ab52. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1244.230672] env[63355]: DEBUG nova.network.neutron [req-6039dfcd-5bd5-4f3f-9f89-24d8e3af73de req-7df3334f-5ed8-4c0f-b183-7119fd801c4c service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updating instance_info_cache with network_info: [{"id": "57106266-129a-4058-8690-6deccdf8ab52", "address": "fa:16:3e:9e:a7:c8", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap57106266-12", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1244.315948] env[63355]: DEBUG nova.scheduler.client.report [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1244.624622] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f0c35d9f-70fa-4658-82f4-b53d6d66fd54 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.881s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.733422] env[63355]: DEBUG oslo_concurrency.lockutils [req-6039dfcd-5bd5-4f3f-9f89-24d8e3af73de req-7df3334f-5ed8-4c0f-b183-7119fd801c4c service nova] Releasing lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1244.823674] env[63355]: DEBUG oslo_concurrency.lockutils [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.125s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.331847] env[63355]: DEBUG oslo_concurrency.lockutils [None req-45100b41-8cfc-4501-a3fa-3532fd14053f tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "500cef8e-f058-4597-8410-ff83bfab06eb" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.090s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.332499] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "500cef8e-f058-4597-8410-ff83bfab06eb" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.491s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1245.333195] env[63355]: INFO nova.compute.manager [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Unshelving [ 1245.544565] env[63355]: DEBUG nova.compute.manager [req-fbbd0362-a25f-411b-8b9b-6428b8ad79ba req-2597b405-67b0-41ce-ac22-2e17a189d455 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Received event network-changed-bc7a8f76-e220-4474-8089-83364388f489 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1245.544758] env[63355]: DEBUG nova.compute.manager [req-fbbd0362-a25f-411b-8b9b-6428b8ad79ba req-2597b405-67b0-41ce-ac22-2e17a189d455 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Refreshing instance network info cache due to event network-changed-bc7a8f76-e220-4474-8089-83364388f489. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1245.544976] env[63355]: DEBUG oslo_concurrency.lockutils [req-fbbd0362-a25f-411b-8b9b-6428b8ad79ba req-2597b405-67b0-41ce-ac22-2e17a189d455 service nova] Acquiring lock "refresh_cache-e4c425fa-0b13-456c-b12b-c1ff18593536" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1245.545136] env[63355]: DEBUG oslo_concurrency.lockutils [req-fbbd0362-a25f-411b-8b9b-6428b8ad79ba req-2597b405-67b0-41ce-ac22-2e17a189d455 service nova] Acquired lock "refresh_cache-e4c425fa-0b13-456c-b12b-c1ff18593536" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1245.545299] env[63355]: DEBUG nova.network.neutron [req-fbbd0362-a25f-411b-8b9b-6428b8ad79ba req-2597b405-67b0-41ce-ac22-2e17a189d455 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Refreshing network info cache for port bc7a8f76-e220-4474-8089-83364388f489 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1246.243195] env[63355]: DEBUG nova.network.neutron [req-fbbd0362-a25f-411b-8b9b-6428b8ad79ba req-2597b405-67b0-41ce-ac22-2e17a189d455 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Updated VIF entry in instance network info cache for port bc7a8f76-e220-4474-8089-83364388f489. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1246.243598] env[63355]: DEBUG nova.network.neutron [req-fbbd0362-a25f-411b-8b9b-6428b8ad79ba req-2597b405-67b0-41ce-ac22-2e17a189d455 service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Updating instance_info_cache with network_info: [{"id": "bc7a8f76-e220-4474-8089-83364388f489", "address": "fa:16:3e:0d:c1:e0", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc7a8f76-e2", "ovs_interfaceid": "bc7a8f76-e220-4474-8089-83364388f489", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1246.359189] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.359487] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1246.359702] env[63355]: DEBUG nova.objects.instance [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'pci_requests' on Instance uuid 500cef8e-f058-4597-8410-ff83bfab06eb {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1246.746125] env[63355]: DEBUG oslo_concurrency.lockutils [req-fbbd0362-a25f-411b-8b9b-6428b8ad79ba req-2597b405-67b0-41ce-ac22-2e17a189d455 service nova] Releasing lock "refresh_cache-e4c425fa-0b13-456c-b12b-c1ff18593536" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1246.863552] env[63355]: DEBUG nova.objects.instance [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'numa_topology' on Instance uuid 500cef8e-f058-4597-8410-ff83bfab06eb {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1247.370546] env[63355]: INFO nova.compute.claims [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1248.430676] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bdf1d7-2aa9-4ec7-b69c-6ba03207d2de {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.437559] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4825573b-44fb-46f2-b363-df1bf1397fc9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.466963] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5effb89-08fd-47c3-8b41-b6f9d12e670b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.473763] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3088740-83b1-4210-ae9b-bcbeb8213e12 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.486559] env[63355]: DEBUG nova.compute.provider_tree [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1248.990045] env[63355]: DEBUG nova.scheduler.client.report [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1249.133212] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1249.496061] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.136s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.524262] env[63355]: INFO nova.network.neutron [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updating port 57106266-129a-4058-8690-6deccdf8ab52 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1249.636234] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.636475] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.636640] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.636797] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1249.637726] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0254b2-f181-4af2-b328-f327ae6f778d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.646152] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ff43e2-30eb-4069-b9c9-2cb3184db4df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.659511] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37576b61-3845-4630-aa47-4d2a75cf09ad {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.665482] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282e5345-1a46-4206-9d2f-6caa93981361 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.695106] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180683MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1249.695295] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.695733] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.717516] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8db23054-4556-496a-a4b1-a24c71c30f50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1250.717829] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 5b246e6a-9974-49e4-b213-7664c90e8420 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1250.717829] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance e4c425fa-0b13-456c-b12b-c1ff18593536 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1250.717906] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 500cef8e-f058-4597-8410-ff83bfab06eb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1250.718089] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1250.718227] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1250.776174] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd68159-4b82-4b9a-a436-01d3c9049d0f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.783907] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1069f71c-4bdd-487a-94c7-85c4f2ca8167 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.813554] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6497b883-9c3e-4943-a298-a022f3ffa2ee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.820626] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0390a906-01be-4618-a7c2-303c528ce709 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.835516] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1250.905042] env[63355]: DEBUG nova.compute.manager [req-917216a8-7220-44ad-816b-35f2870853e6 req-3d1a4e0e-dc7d-4024-949f-ea6cbe2dd602 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received event network-vif-plugged-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1250.905267] env[63355]: DEBUG oslo_concurrency.lockutils [req-917216a8-7220-44ad-816b-35f2870853e6 req-3d1a4e0e-dc7d-4024-949f-ea6cbe2dd602 service nova] Acquiring lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.905479] env[63355]: DEBUG oslo_concurrency.lockutils [req-917216a8-7220-44ad-816b-35f2870853e6 req-3d1a4e0e-dc7d-4024-949f-ea6cbe2dd602 service nova] Lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.905652] env[63355]: DEBUG oslo_concurrency.lockutils [req-917216a8-7220-44ad-816b-35f2870853e6 req-3d1a4e0e-dc7d-4024-949f-ea6cbe2dd602 service nova] Lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1250.905817] env[63355]: DEBUG nova.compute.manager [req-917216a8-7220-44ad-816b-35f2870853e6 req-3d1a4e0e-dc7d-4024-949f-ea6cbe2dd602 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] No waiting events found dispatching network-vif-plugged-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1250.905982] env[63355]: WARNING nova.compute.manager [req-917216a8-7220-44ad-816b-35f2870853e6 req-3d1a4e0e-dc7d-4024-949f-ea6cbe2dd602 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received unexpected event network-vif-plugged-57106266-129a-4058-8690-6deccdf8ab52 for instance with vm_state shelved_offloaded and task_state spawning. [ 1250.989799] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.989799] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.989799] env[63355]: DEBUG nova.network.neutron [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1251.338172] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1251.688146] env[63355]: DEBUG nova.network.neutron [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updating instance_info_cache with network_info: [{"id": "57106266-129a-4058-8690-6deccdf8ab52", "address": "fa:16:3e:9e:a7:c8", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57106266-12", "ovs_interfaceid": "57106266-129a-4058-8690-6deccdf8ab52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1251.843186] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1251.843531] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.148s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.191400] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1252.217267] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='fd677575ff26b72e6a145818c458872e',container_format='bare',created_at=2024-10-24T02:35:30Z,direct_url=,disk_format='vmdk',id=f25b0300-4805-4fd8-9157-fefa86c07e1a,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-492943034-shelved',owner='6b3fee9bc99d49ea9de53d5dce52c79d',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-24T02:35:42Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1252.217536] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1252.217705] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1252.217919] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1252.218093] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1252.218249] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1252.218459] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1252.218623] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1252.218792] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1252.218956] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1252.219142] env[63355]: DEBUG nova.virt.hardware [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1252.220052] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c646f0e-d28e-46a5-b92f-7d1cc99df5de {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.228201] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc229c4-c0e7-43fa-8887-14701bc32898 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.241100] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:a7:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57106266-129a-4058-8690-6deccdf8ab52', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1252.248257] env[63355]: DEBUG oslo.service.loopingcall [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1252.248490] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1252.248690] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c6032b6-cf87-4cff-b2fc-6f861d06ab20 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.267571] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1252.267571] env[63355]: value = "task-1350306" [ 1252.267571] env[63355]: _type = "Task" [ 1252.267571] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.274908] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350306, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.777088] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350306, 'name': CreateVM_Task, 'duration_secs': 0.345035} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.777269] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1252.777956] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1252.778148] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.778532] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1252.778787] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31525776-baf4-445c-9d0e-6a03895bde1e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.782932] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1252.782932] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52db4c72-d82e-5e5c-8974-6261b98c8601" [ 1252.782932] env[63355]: _type = "Task" [ 1252.782932] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.789690] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52db4c72-d82e-5e5c-8974-6261b98c8601, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.929649] env[63355]: DEBUG nova.compute.manager [req-59072bd8-3eae-4901-9003-e9eb71405bad req-d1ec7e98-4055-4c83-aa3a-c0486265931d service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received event network-changed-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1252.929859] env[63355]: DEBUG nova.compute.manager [req-59072bd8-3eae-4901-9003-e9eb71405bad req-d1ec7e98-4055-4c83-aa3a-c0486265931d service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Refreshing instance network info cache due to event network-changed-57106266-129a-4058-8690-6deccdf8ab52. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1252.930012] env[63355]: DEBUG oslo_concurrency.lockutils [req-59072bd8-3eae-4901-9003-e9eb71405bad req-d1ec7e98-4055-4c83-aa3a-c0486265931d service nova] Acquiring lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1252.930175] env[63355]: DEBUG oslo_concurrency.lockutils [req-59072bd8-3eae-4901-9003-e9eb71405bad req-d1ec7e98-4055-4c83-aa3a-c0486265931d service nova] Acquired lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.930338] env[63355]: DEBUG nova.network.neutron [req-59072bd8-3eae-4901-9003-e9eb71405bad req-d1ec7e98-4055-4c83-aa3a-c0486265931d service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Refreshing network info cache for port 57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1253.292640] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1253.292905] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Processing image f25b0300-4805-4fd8-9157-fefa86c07e1a {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1253.293169] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a/f25b0300-4805-4fd8-9157-fefa86c07e1a.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1253.293323] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a/f25b0300-4805-4fd8-9157-fefa86c07e1a.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1253.293508] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1253.293761] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8b96f5b-f1bd-4121-955d-f5555e1c37f8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.301930] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1253.302115] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1253.302760] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afb0e2f1-5307-4ba4-9995-dd7f562c4965 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.307189] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1253.307189] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5203bcaa-23ef-916f-b2d6-9e4689f3aeee" [ 1253.307189] env[63355]: _type = "Task" [ 1253.307189] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.315545] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5203bcaa-23ef-916f-b2d6-9e4689f3aeee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.622103] env[63355]: DEBUG nova.network.neutron [req-59072bd8-3eae-4901-9003-e9eb71405bad req-d1ec7e98-4055-4c83-aa3a-c0486265931d service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updated VIF entry in instance network info cache for port 57106266-129a-4058-8690-6deccdf8ab52. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1253.622505] env[63355]: DEBUG nova.network.neutron [req-59072bd8-3eae-4901-9003-e9eb71405bad req-d1ec7e98-4055-4c83-aa3a-c0486265931d service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updating instance_info_cache with network_info: [{"id": "57106266-129a-4058-8690-6deccdf8ab52", "address": "fa:16:3e:9e:a7:c8", "network": {"id": "7fd03f9f-3853-4ca6-8fe4-099318f8785d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1932757459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3fee9bc99d49ea9de53d5dce52c79d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57106266-12", "ovs_interfaceid": "57106266-129a-4058-8690-6deccdf8ab52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1253.817883] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Preparing fetch location {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1253.817883] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Fetch image to [datastore1] OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6/OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6.vmdk {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1253.817883] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Downloading stream optimized image f25b0300-4805-4fd8-9157-fefa86c07e1a to [datastore1] OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6/OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6.vmdk on the data store datastore1 as vApp {{(pid=63355) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1253.817883] env[63355]: DEBUG nova.virt.vmwareapi.images [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Downloading image file data f25b0300-4805-4fd8-9157-fefa86c07e1a to the ESX as VM named 'OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6' {{(pid=63355) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1253.880939] env[63355]: DEBUG oslo_vmware.rw_handles [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1253.880939] env[63355]: value = "resgroup-9" [ 1253.880939] env[63355]: _type = "ResourcePool" [ 1253.880939] env[63355]: }. {{(pid=63355) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1253.881223] env[63355]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-f009a277-f466-4eac-8cfe-6789b3a0d2cc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.902659] env[63355]: DEBUG oslo_vmware.rw_handles [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lease: (returnval){ [ 1253.902659] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a05fc6-1b61-d8cf-206d-a09087b7be79" [ 1253.902659] env[63355]: _type = "HttpNfcLease" [ 1253.902659] env[63355]: } obtained for vApp import into resource pool (val){ [ 1253.902659] env[63355]: value = "resgroup-9" [ 1253.902659] env[63355]: _type = "ResourcePool" [ 1253.902659] env[63355]: }. {{(pid=63355) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1253.903070] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the lease: (returnval){ [ 1253.903070] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a05fc6-1b61-d8cf-206d-a09087b7be79" [ 1253.903070] env[63355]: _type = "HttpNfcLease" [ 1253.903070] env[63355]: } to be ready. {{(pid=63355) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1253.908707] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1253.908707] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a05fc6-1b61-d8cf-206d-a09087b7be79" [ 1253.908707] env[63355]: _type = "HttpNfcLease" [ 1253.908707] env[63355]: } is initializing. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1254.125542] env[63355]: DEBUG oslo_concurrency.lockutils [req-59072bd8-3eae-4901-9003-e9eb71405bad req-d1ec7e98-4055-4c83-aa3a-c0486265931d service nova] Releasing lock "refresh_cache-500cef8e-f058-4597-8410-ff83bfab06eb" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1254.410951] env[63355]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1254.410951] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a05fc6-1b61-d8cf-206d-a09087b7be79" [ 1254.410951] env[63355]: _type = "HttpNfcLease" [ 1254.410951] env[63355]: } is ready. {{(pid=63355) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1254.411259] env[63355]: DEBUG oslo_vmware.rw_handles [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1254.411259] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52a05fc6-1b61-d8cf-206d-a09087b7be79" [ 1254.411259] env[63355]: _type = "HttpNfcLease" [ 1254.411259] env[63355]: }. {{(pid=63355) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1254.411919] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5babfe3e-2c65-42b1-bd6d-46e078c41ec0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.418761] env[63355]: DEBUG oslo_vmware.rw_handles [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c64d9d-a4b0-5dd5-af0b-d828d0aa8208/disk-0.vmdk from lease info. {{(pid=63355) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1254.418937] env[63355]: DEBUG oslo_vmware.rw_handles [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c64d9d-a4b0-5dd5-af0b-d828d0aa8208/disk-0.vmdk. {{(pid=63355) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1254.481496] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e23b47e2-a25e-4bc1-bb4a-4d4f0709e33f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.621544] env[63355]: DEBUG oslo_vmware.rw_handles [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Completed reading data from the image iterator. {{(pid=63355) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1255.621904] env[63355]: DEBUG oslo_vmware.rw_handles [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c64d9d-a4b0-5dd5-af0b-d828d0aa8208/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1255.622745] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350daadf-7021-4b60-ac04-45da9d608b6f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.629371] env[63355]: DEBUG oslo_vmware.rw_handles [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c64d9d-a4b0-5dd5-af0b-d828d0aa8208/disk-0.vmdk is in state: ready. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1255.629596] env[63355]: DEBUG oslo_vmware.rw_handles [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c64d9d-a4b0-5dd5-af0b-d828d0aa8208/disk-0.vmdk. {{(pid=63355) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1255.629832] env[63355]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-d7b2618e-5216-4fbb-a7b5-c8280c4e437f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.820329] env[63355]: DEBUG oslo_vmware.rw_handles [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c64d9d-a4b0-5dd5-af0b-d828d0aa8208/disk-0.vmdk. {{(pid=63355) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1255.820554] env[63355]: INFO nova.virt.vmwareapi.images [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Downloaded image file data f25b0300-4805-4fd8-9157-fefa86c07e1a [ 1255.821435] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0768330f-d39d-435c-954d-37c02ff6b4a7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.836128] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-767c4403-ced3-4051-94f2-81df0dc2d778 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.839130] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.839317] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.839484] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1255.871105] env[63355]: INFO nova.virt.vmwareapi.images [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] The imported VM was unregistered [ 1255.873439] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Caching image {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1255.873679] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Creating directory with path [datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1255.874177] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c72f2af-9cb4-4ef8-ad4f-9630366fbe53 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.896624] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Created directory with path [datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1255.896814] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6/OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6.vmdk to [datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a/f25b0300-4805-4fd8-9157-fefa86c07e1a.vmdk. {{(pid=63355) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1255.897085] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-487a21b5-0bd0-417c-a41d-153ee5b050ab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.904465] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1255.904465] env[63355]: value = "task-1350309" [ 1255.904465] env[63355]: _type = "Task" [ 1255.904465] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.912253] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350309, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.209138] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "5b246e6a-9974-49e4-b213-7664c90e8420" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1256.209389] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1256.382141] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.382303] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquired lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.382476] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Forcefully refreshing network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1256.416231] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350309, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.712874] env[63355]: DEBUG nova.compute.utils [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1256.915119] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350309, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.216052] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.416240] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350309, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.642466] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updating instance_info_cache with network_info: [{"id": "e3455ba6-ed91-4901-b232-76ca673421a7", "address": "fa:16:3e:eb:3b:b7", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3455ba6-ed", "ovs_interfaceid": "e3455ba6-ed91-4901-b232-76ca673421a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1257.916360] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350309, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.145172] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Releasing lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.145444] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updated the network info_cache for instance {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1258.145664] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.145822] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.145972] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.146143] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.146311] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.146486] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.146603] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1258.283967] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "5b246e6a-9974-49e4-b213-7664c90e8420" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1258.284266] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1258.284506] env[63355]: INFO nova.compute.manager [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Attaching volume 6f56474d-ef78-42b3-974e-5b060951acb4 to /dev/sdb [ 1258.328372] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c986647a-5e15-4181-865b-e4a924422204 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.335562] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109f1203-4c1c-41f4-83be-6559b04cfb57 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.349008] env[63355]: DEBUG nova.virt.block_device [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Updating existing volume attachment record: 7f09acb1-54a7-4b37-b44e-9a9ab5328cf7 {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1258.418204] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350309, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.415865} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.418506] env[63355]: INFO nova.virt.vmwareapi.ds_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6/OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6.vmdk to [datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a/f25b0300-4805-4fd8-9157-fefa86c07e1a.vmdk. [ 1258.418708] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Cleaning up location [datastore1] OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6 {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1258.418880] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_aaa3fdec-cb20-469b-8246-da9a9b5ccda6 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1258.419153] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3f0192e-3370-4419-acf8-aa0e949cc2c3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.426033] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1258.426033] env[63355]: value = "task-1350310" [ 1258.426033] env[63355]: _type = "Task" [ 1258.426033] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.432264] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350310, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.934827] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350310, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.032063} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.935310] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1258.935310] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a/f25b0300-4805-4fd8-9157-fefa86c07e1a.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.935513] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a/f25b0300-4805-4fd8-9157-fefa86c07e1a.vmdk to [datastore1] 500cef8e-f058-4597-8410-ff83bfab06eb/500cef8e-f058-4597-8410-ff83bfab06eb.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1258.935786] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f51267c2-952a-4d22-9bf7-25fd1e4faf34 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.941614] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1258.941614] env[63355]: value = "task-1350314" [ 1258.941614] env[63355]: _type = "Task" [ 1258.941614] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.950469] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350314, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.452185] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350314, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.955454] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350314, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.453916] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350314, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.953896] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350314, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.454671] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350314, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.382656} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.455081] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f25b0300-4805-4fd8-9157-fefa86c07e1a/f25b0300-4805-4fd8-9157-fefa86c07e1a.vmdk to [datastore1] 500cef8e-f058-4597-8410-ff83bfab06eb/500cef8e-f058-4597-8410-ff83bfab06eb.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1261.455695] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d124ccf-0a65-49c7-8d75-5ce4a1a499ae {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.476267] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 500cef8e-f058-4597-8410-ff83bfab06eb/500cef8e-f058-4597-8410-ff83bfab06eb.vmdk or device None with type streamOptimized {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1261.476489] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8db52660-1fc6-47a4-b5f0-6390787790cd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.494587] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1261.494587] env[63355]: value = "task-1350316" [ 1261.494587] env[63355]: _type = "Task" [ 1261.494587] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.501488] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350316, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.004056] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350316, 'name': ReconfigVM_Task, 'duration_secs': 0.252999} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.004347] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 500cef8e-f058-4597-8410-ff83bfab06eb/500cef8e-f058-4597-8410-ff83bfab06eb.vmdk or device None with type streamOptimized {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1262.004966] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80cb6177-382b-4776-88fa-91ff4d37afe5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.010570] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1262.010570] env[63355]: value = "task-1350317" [ 1262.010570] env[63355]: _type = "Task" [ 1262.010570] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.018579] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350317, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.520255] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350317, 'name': Rename_Task, 'duration_secs': 0.129928} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.520607] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1262.520712] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9622a47e-97f8-4166-93e2-06cc9fbe127b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.527291] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1262.527291] env[63355]: value = "task-1350318" [ 1262.527291] env[63355]: _type = "Task" [ 1262.527291] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.534276] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350318, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.892699] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1262.893037] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287781', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'name': 'volume-6f56474d-ef78-42b3-974e-5b060951acb4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b246e6a-9974-49e4-b213-7664c90e8420', 'attached_at': '', 'detached_at': '', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'serial': '6f56474d-ef78-42b3-974e-5b060951acb4'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1262.894074] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92fddd0b-ea94-4ca5-8bfd-135337678fd5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.912144] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f463aba-62bb-4d23-a056-36fef3128c74 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.938639] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-6f56474d-ef78-42b3-974e-5b060951acb4/volume-6f56474d-ef78-42b3-974e-5b060951acb4.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1262.938957] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28551df0-c221-40a5-aca9-8d077c3c2754 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.956939] env[63355]: DEBUG oslo_vmware.api [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1262.956939] env[63355]: value = "task-1350319" [ 1262.956939] env[63355]: _type = "Task" [ 1262.956939] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.965888] env[63355]: DEBUG oslo_vmware.api [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350319, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.037400] env[63355]: DEBUG oslo_vmware.api [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350318, 'name': PowerOnVM_Task, 'duration_secs': 0.442348} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.037734] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1263.137264] env[63355]: DEBUG nova.compute.manager [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1263.138211] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18f4c79-e0aa-4370-bbc4-b7bf81b40c9f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.466200] env[63355]: DEBUG oslo_vmware.api [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350319, 'name': ReconfigVM_Task, 'duration_secs': 0.335109} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.466479] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-6f56474d-ef78-42b3-974e-5b060951acb4/volume-6f56474d-ef78-42b3-974e-5b060951acb4.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1263.471030] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-797af1a2-2677-4289-bf04-e4d0af7dc693 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.485224] env[63355]: DEBUG oslo_vmware.api [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1263.485224] env[63355]: value = "task-1350320" [ 1263.485224] env[63355]: _type = "Task" [ 1263.485224] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.492406] env[63355]: DEBUG oslo_vmware.api [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350320, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.656878] env[63355]: DEBUG oslo_concurrency.lockutils [None req-0b059a0a-1f2b-4b41-8cd0-303c40910453 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "500cef8e-f058-4597-8410-ff83bfab06eb" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.324s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1263.995298] env[63355]: DEBUG oslo_vmware.api [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350320, 'name': ReconfigVM_Task, 'duration_secs': 0.15024} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.995619] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287781', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'name': 'volume-6f56474d-ef78-42b3-974e-5b060951acb4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b246e6a-9974-49e4-b213-7664c90e8420', 'attached_at': '', 'detached_at': '', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'serial': '6f56474d-ef78-42b3-974e-5b060951acb4'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1264.503993] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "500cef8e-f058-4597-8410-ff83bfab06eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.504294] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "500cef8e-f058-4597-8410-ff83bfab06eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.504688] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.504688] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.504854] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "500cef8e-f058-4597-8410-ff83bfab06eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.509030] env[63355]: INFO nova.compute.manager [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Terminating instance [ 1264.510610] env[63355]: DEBUG nova.compute.manager [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1264.510808] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1264.511656] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e467ac-a46f-4ec0-a362-bef56bb8c5da {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.519741] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1264.519945] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50ebaf86-be77-4aa0-8324-55f324b6beb5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.526953] env[63355]: DEBUG oslo_vmware.api [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1264.526953] env[63355]: value = "task-1350321" [ 1264.526953] env[63355]: _type = "Task" [ 1264.526953] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.536878] env[63355]: DEBUG oslo_vmware.api [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350321, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.037301] env[63355]: DEBUG oslo_vmware.api [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350321, 'name': PowerOffVM_Task, 'duration_secs': 0.156766} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.038137] env[63355]: DEBUG nova.objects.instance [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lazy-loading 'flavor' on Instance uuid 5b246e6a-9974-49e4-b213-7664c90e8420 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1265.040028] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1265.040028] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1265.040171] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ccfc66bb-028a-46bd-9439-d30ace0fd5fd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.098831] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1265.099095] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1265.099322] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleting the datastore file [datastore1] 500cef8e-f058-4597-8410-ff83bfab06eb {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1265.099649] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f31747be-3a13-443e-99f3-5a342c8b8170 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.105450] env[63355]: DEBUG oslo_vmware.api [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for the task: (returnval){ [ 1265.105450] env[63355]: value = "task-1350323" [ 1265.105450] env[63355]: _type = "Task" [ 1265.105450] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.113552] env[63355]: DEBUG oslo_vmware.api [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350323, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.543239] env[63355]: DEBUG oslo_concurrency.lockutils [None req-01524674-9cd0-49fe-b15c-9657f16c44d7 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.259s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1265.615552] env[63355]: DEBUG oslo_vmware.api [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Task: {'id': task-1350323, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157337} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.615822] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1265.616063] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1265.616360] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1265.616654] env[63355]: INFO nova.compute.manager [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1265.617036] env[63355]: DEBUG oslo.service.loopingcall [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1265.617325] env[63355]: DEBUG nova.compute.manager [-] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1265.617477] env[63355]: DEBUG nova.network.neutron [-] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1265.922470] env[63355]: INFO nova.compute.manager [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Rebuilding instance [ 1265.970057] env[63355]: DEBUG nova.compute.manager [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1265.970862] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0b0571-96d9-42c5-bd54-ad7721fdc368 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.065556] env[63355]: DEBUG nova.compute.manager [req-e8609c85-f529-4ca3-bcd1-b344dc1305db req-5afbd5df-5f1d-4c1d-bdeb-fc58dc891592 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Received event network-vif-deleted-57106266-129a-4058-8690-6deccdf8ab52 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1266.065809] env[63355]: INFO nova.compute.manager [req-e8609c85-f529-4ca3-bcd1-b344dc1305db req-5afbd5df-5f1d-4c1d-bdeb-fc58dc891592 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Neutron deleted interface 57106266-129a-4058-8690-6deccdf8ab52; detaching it from the instance and deleting it from the info cache [ 1266.065942] env[63355]: DEBUG nova.network.neutron [req-e8609c85-f529-4ca3-bcd1-b344dc1305db req-5afbd5df-5f1d-4c1d-bdeb-fc58dc891592 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1266.435703] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.482596] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1266.483097] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8ccc5ab1-6770-442f-8198-4758dbf1c43b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.491123] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1266.491123] env[63355]: value = "task-1350324" [ 1266.491123] env[63355]: _type = "Task" [ 1266.491123] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.499641] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350324, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.543731] env[63355]: DEBUG nova.network.neutron [-] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1266.568113] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d25ff348-01ec-4899-9b8b-8eaaeca2cfc0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.580176] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0560a8f8-ef18-400b-8ed8-a3271dcca101 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.608050] env[63355]: DEBUG nova.compute.manager [req-e8609c85-f529-4ca3-bcd1-b344dc1305db req-5afbd5df-5f1d-4c1d-bdeb-fc58dc891592 service nova] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Detach interface failed, port_id=57106266-129a-4058-8690-6deccdf8ab52, reason: Instance 500cef8e-f058-4597-8410-ff83bfab06eb could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1267.001654] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350324, 'name': PowerOffVM_Task, 'duration_secs': 0.186964} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.001920] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1267.046254] env[63355]: INFO nova.compute.manager [-] [instance: 500cef8e-f058-4597-8410-ff83bfab06eb] Took 1.43 seconds to deallocate network for instance. [ 1267.056741] env[63355]: INFO nova.compute.manager [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Detaching volume 6f56474d-ef78-42b3-974e-5b060951acb4 [ 1267.096841] env[63355]: INFO nova.virt.block_device [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Attempting to driver detach volume 6f56474d-ef78-42b3-974e-5b060951acb4 from mountpoint /dev/sdb [ 1267.097102] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1267.097438] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287781', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'name': 'volume-6f56474d-ef78-42b3-974e-5b060951acb4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b246e6a-9974-49e4-b213-7664c90e8420', 'attached_at': '', 'detached_at': '', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'serial': '6f56474d-ef78-42b3-974e-5b060951acb4'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1267.098412] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771c9d61-6a98-4834-9cc1-f74c2e39a352 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.119913] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4282a3e2-3037-4315-89ef-8fad7422f2e9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.126545] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3fa61f-6a6d-4b85-8380-6632924241ca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.146907] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c4ac4b-c565-4e96-b867-b847a088eec7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.161197] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] The volume has not been displaced from its original location: [datastore2] volume-6f56474d-ef78-42b3-974e-5b060951acb4/volume-6f56474d-ef78-42b3-974e-5b060951acb4.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1267.166479] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1267.166742] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62fff309-7bb7-437d-9ddd-54eaf7ad06fd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.185125] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1267.185125] env[63355]: value = "task-1350325" [ 1267.185125] env[63355]: _type = "Task" [ 1267.185125] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.195094] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350325, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.557245] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1267.557532] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1267.557760] env[63355]: DEBUG nova.objects.instance [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lazy-loading 'resources' on Instance uuid 500cef8e-f058-4597-8410-ff83bfab06eb {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1267.694358] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350325, 'name': ReconfigVM_Task, 'duration_secs': 0.192988} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.694641] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1267.699386] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12f64dab-03d6-4681-b6da-14bc7211d8ae {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.713744] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1267.713744] env[63355]: value = "task-1350326" [ 1267.713744] env[63355]: _type = "Task" [ 1267.713744] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.721515] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350326, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.121447] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd61fea3-7041-4f5b-a74a-e311643c2579 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.129261] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2746184-f7d1-4ad4-8512-8c98f545feb2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.157832] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e4f96b-ec54-44cf-873a-6d295014d26c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.164552] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1acdfa6-5515-4e66-80e5-7eff5f8af530 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.177050] env[63355]: DEBUG nova.compute.provider_tree [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1268.223434] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350326, 'name': ReconfigVM_Task, 'duration_secs': 0.175239} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.223707] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287781', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'name': 'volume-6f56474d-ef78-42b3-974e-5b060951acb4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b246e6a-9974-49e4-b213-7664c90e8420', 'attached_at': '', 'detached_at': '', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'serial': '6f56474d-ef78-42b3-974e-5b060951acb4'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1268.680450] env[63355]: DEBUG nova.scheduler.client.report [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1269.185317] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.628s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.208147] env[63355]: INFO nova.scheduler.client.report [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Deleted allocations for instance 500cef8e-f058-4597-8410-ff83bfab06eb [ 1269.269791] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1269.270109] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca4de0c7-e6ee-4d80-8409-59af0f7a8177 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.277258] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1269.277258] env[63355]: value = "task-1350327" [ 1269.277258] env[63355]: _type = "Task" [ 1269.277258] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.717667] env[63355]: DEBUG oslo_concurrency.lockutils [None req-bc33ad29-64c9-4c78-8009-493d12d3ce69 tempest-ServerActionsTestOtherB-1959432081 tempest-ServerActionsTestOtherB-1959432081-project-member] Lock "500cef8e-f058-4597-8410-ff83bfab06eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.213s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.787391] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] VM already powered off {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1269.787582] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1269.787781] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287781', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'name': 'volume-6f56474d-ef78-42b3-974e-5b060951acb4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b246e6a-9974-49e4-b213-7664c90e8420', 'attached_at': '', 'detached_at': '', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'serial': '6f56474d-ef78-42b3-974e-5b060951acb4'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1269.788589] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103ce928-901d-42d5-adcd-da6a139d7690 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.806035] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a2ffd00-5d32-4242-abfe-6f9062b5f915 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.811424] env[63355]: WARNING nova.virt.vmwareapi.driver [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1269.811674] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1269.812376] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a72324-a2e2-437d-a5e7-7877e0a0b701 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.818577] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1269.818798] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-160c6639-b7b9-4807-96e8-a5282e401017 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.890781] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1269.891018] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1269.891225] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleting the datastore file [datastore1] 5b246e6a-9974-49e4-b213-7664c90e8420 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1269.891487] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74298cf6-81ee-4539-af33-0f0d51d177a2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.897988] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1269.897988] env[63355]: value = "task-1350330" [ 1269.897988] env[63355]: _type = "Task" [ 1269.897988] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.905199] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.408587] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138173} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.408587] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1270.408587] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1270.408587] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1270.913647] env[63355]: INFO nova.virt.block_device [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Booting with volume 6f56474d-ef78-42b3-974e-5b060951acb4 at /dev/sdb [ 1270.946665] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8ed6793-cbb0-42c7-90a1-bbd4382f3b32 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.956147] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8fa61c-aab3-44fa-886d-070ebbd072bc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.980612] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-acae87b3-5c99-4f26-b201-dc90b854b8fb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.988801] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd969b0e-054d-4a4e-a043-6dbd8fb609b9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.013370] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1061a288-f372-4155-b4bb-5f69425a3f84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.019662] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9e5961-b76d-4be3-886c-34730f873bb2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.032955] env[63355]: DEBUG nova.virt.block_device [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Updating existing volume attachment record: b9e7f52e-c245-41ab-9766-e7ba5dc77e5c {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1273.149750] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1273.150099] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1273.150203] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1273.150391] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1273.150716] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1273.150716] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1273.150898] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1273.151152] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1273.151351] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1273.151515] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1273.151690] env[63355]: DEBUG nova.virt.hardware [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1273.152672] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e17af61-763c-421b-a544-f0412a81374d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.160362] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e643ab-ad63-4e35-84a4-0125dbf64adb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.174593] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:e5:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '313f0c1d-9da1-41ec-bba9-5ff65d52db63', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1273.182220] env[63355]: DEBUG oslo.service.loopingcall [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1273.182472] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1273.182692] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-07a54ff2-8aee-483a-a963-05227440ea7f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.201767] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1273.201767] env[63355]: value = "task-1350331" [ 1273.201767] env[63355]: _type = "Task" [ 1273.201767] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.210330] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350331, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.712235] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350331, 'name': CreateVM_Task, 'duration_secs': 0.354576} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.712441] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1273.713131] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1273.713313] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1273.713721] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1273.714222] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3830b0e7-5a4e-4655-9630-d00d744bd8b4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.718435] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1273.718435] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52959751-bf97-b76d-8ef2-160328380cba" [ 1273.718435] env[63355]: _type = "Task" [ 1273.718435] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.726039] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52959751-bf97-b76d-8ef2-160328380cba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.229880] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52959751-bf97-b76d-8ef2-160328380cba, 'name': SearchDatastore_Task, 'duration_secs': 0.009096} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.230225] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1274.230441] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1274.230653] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1274.230831] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.231033] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1274.231310] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22a1e9af-e681-42fe-8ab5-2dd853743d0e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.239879] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1274.240068] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1274.240817] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0556a28-3d01-4571-baed-54ac1a937b6e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.247257] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1274.247257] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5280172f-93b6-2eaa-3596-6be5ca80613d" [ 1274.247257] env[63355]: _type = "Task" [ 1274.247257] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.255014] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5280172f-93b6-2eaa-3596-6be5ca80613d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.757411] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5280172f-93b6-2eaa-3596-6be5ca80613d, 'name': SearchDatastore_Task, 'duration_secs': 0.008368} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.758214] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c73e76de-5b71-4bf3-a3f6-71e99b9bb797 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.763257] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1274.763257] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52267b17-72ed-e0ee-1a4f-4bbfb88c8997" [ 1274.763257] env[63355]: _type = "Task" [ 1274.763257] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.770665] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52267b17-72ed-e0ee-1a4f-4bbfb88c8997, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.274224] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52267b17-72ed-e0ee-1a4f-4bbfb88c8997, 'name': SearchDatastore_Task, 'duration_secs': 0.008498} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.274592] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1275.274818] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 5b246e6a-9974-49e4-b213-7664c90e8420/5b246e6a-9974-49e4-b213-7664c90e8420.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1275.275132] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b74ea41-83f7-42be-8563-1067fcf28c45 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.281361] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1275.281361] env[63355]: value = "task-1350332" [ 1275.281361] env[63355]: _type = "Task" [ 1275.281361] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.288414] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350332, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.790622] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350332, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490492} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.791039] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 5b246e6a-9974-49e4-b213-7664c90e8420/5b246e6a-9974-49e4-b213-7664c90e8420.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1275.791295] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1275.791594] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-31088930-1c23-4d16-ad14-2cc3f9cb8449 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.797564] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1275.797564] env[63355]: value = "task-1350333" [ 1275.797564] env[63355]: _type = "Task" [ 1275.797564] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.804702] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350333, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.307389] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350333, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061925} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.307748] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1276.308412] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e13824-4d8c-47ac-9853-1e2fed391355 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.329299] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 5b246e6a-9974-49e4-b213-7664c90e8420/5b246e6a-9974-49e4-b213-7664c90e8420.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1276.329532] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59033198-7072-44cc-b81e-d1e437d04aa7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.348086] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1276.348086] env[63355]: value = "task-1350334" [ 1276.348086] env[63355]: _type = "Task" [ 1276.348086] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.354971] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350334, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.858783] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350334, 'name': ReconfigVM_Task, 'duration_secs': 0.251194} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.859159] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 5b246e6a-9974-49e4-b213-7664c90e8420/5b246e6a-9974-49e4-b213-7664c90e8420.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1276.860563] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_options': None, 'boot_index': 0, 'device_name': '/dev/sda', 'size': 0, 'encrypted': False, 'encryption_format': None, 'guest_format': None, 'encryption_secret_uuid': None, 'disk_bus': None, 'device_type': 'disk', 'image_id': '84ca0bb3-9916-4cea-9399-75af77b0558e'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'mount_device': '/dev/sdb', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287781', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'name': 'volume-6f56474d-ef78-42b3-974e-5b060951acb4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b246e6a-9974-49e4-b213-7664c90e8420', 'attached_at': '', 'detached_at': '', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'serial': '6f56474d-ef78-42b3-974e-5b060951acb4'}, 'attachment_id': 'b9e7f52e-c245-41ab-9766-e7ba5dc77e5c', 'guest_format': None, 'delete_on_termination': False, 'disk_bus': None, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=63355) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1276.860785] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1276.860976] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287781', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'name': 'volume-6f56474d-ef78-42b3-974e-5b060951acb4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b246e6a-9974-49e4-b213-7664c90e8420', 'attached_at': '', 'detached_at': '', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'serial': '6f56474d-ef78-42b3-974e-5b060951acb4'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1276.861735] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ebee17-196a-46c9-97e0-cd33e5fdcf7e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.876416] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a2395f6-c95c-426c-8ab5-229febeb604a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.900472] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-6f56474d-ef78-42b3-974e-5b060951acb4/volume-6f56474d-ef78-42b3-974e-5b060951acb4.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1276.900727] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-778c909f-a534-4051-8aab-d50bfa3438fc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.917618] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1276.917618] env[63355]: value = "task-1350335" [ 1276.917618] env[63355]: _type = "Task" [ 1276.917618] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.924690] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350335, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.427117] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350335, 'name': ReconfigVM_Task, 'duration_secs': 0.279253} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.427546] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-6f56474d-ef78-42b3-974e-5b060951acb4/volume-6f56474d-ef78-42b3-974e-5b060951acb4.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1277.432098] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67103240-6d6e-4520-b36a-ad82cd60665a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.446853] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1277.446853] env[63355]: value = "task-1350336" [ 1277.446853] env[63355]: _type = "Task" [ 1277.446853] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.454671] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350336, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.956689] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350336, 'name': ReconfigVM_Task, 'duration_secs': 0.137978} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.956943] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287781', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'name': 'volume-6f56474d-ef78-42b3-974e-5b060951acb4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b246e6a-9974-49e4-b213-7664c90e8420', 'attached_at': '', 'detached_at': '', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'serial': '6f56474d-ef78-42b3-974e-5b060951acb4'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1277.957568] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6875ece1-5aa1-4cb7-8a05-f294b7395017 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.963655] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1277.963655] env[63355]: value = "task-1350337" [ 1277.963655] env[63355]: _type = "Task" [ 1277.963655] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.970661] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350337, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.473538] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350337, 'name': Rename_Task, 'duration_secs': 0.134949} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.473902] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1278.474067] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c4d45c2-6899-4a39-8a5f-a3d93c0fa8df {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.480176] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1278.480176] env[63355]: value = "task-1350338" [ 1278.480176] env[63355]: _type = "Task" [ 1278.480176] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.487162] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350338, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.991218] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350338, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.490438] env[63355]: DEBUG oslo_vmware.api [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350338, 'name': PowerOnVM_Task, 'duration_secs': 0.541708} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.490837] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1279.490920] env[63355]: DEBUG nova.compute.manager [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1279.491695] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3879ce3-f0cc-44c3-aa25-d2cc2a2ad2a2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.009491] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1280.009491] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1280.009491] env[63355]: DEBUG nova.objects.instance [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63355) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1280.856894] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "5b246e6a-9974-49e4-b213-7664c90e8420" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1280.857223] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.017802] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d94f2166-f4ab-4324-8098-c29652271c7f tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.360575] env[63355]: INFO nova.compute.manager [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Detaching volume 6f56474d-ef78-42b3-974e-5b060951acb4 [ 1281.391318] env[63355]: INFO nova.virt.block_device [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Attempting to driver detach volume 6f56474d-ef78-42b3-974e-5b060951acb4 from mountpoint /dev/sdb [ 1281.391633] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1281.391823] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287781', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'name': 'volume-6f56474d-ef78-42b3-974e-5b060951acb4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b246e6a-9974-49e4-b213-7664c90e8420', 'attached_at': '', 'detached_at': '', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'serial': '6f56474d-ef78-42b3-974e-5b060951acb4'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1281.392751] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-005cc1d9-9736-4433-a842-f60aa9abaabf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.414591] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21be311-ff64-4581-9539-b44221a71415 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.421168] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20119ec2-ab4f-405b-86a5-27cedd35a914 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.440336] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5173c16-e41d-4d1d-9125-99f0a081ee64 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.454428] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] The volume has not been displaced from its original location: [datastore2] volume-6f56474d-ef78-42b3-974e-5b060951acb4/volume-6f56474d-ef78-42b3-974e-5b060951acb4.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1281.459479] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1281.459770] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02e50a3b-d74d-43c2-ab2f-3f98cd7b7c31 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.476592] env[63355]: DEBUG oslo_vmware.api [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1281.476592] env[63355]: value = "task-1350339" [ 1281.476592] env[63355]: _type = "Task" [ 1281.476592] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.486181] env[63355]: DEBUG oslo_vmware.api [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350339, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.986684] env[63355]: DEBUG oslo_vmware.api [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350339, 'name': ReconfigVM_Task, 'duration_secs': 0.213682} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.987082] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1281.991614] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4baee212-ca5a-4fb7-9100-614ef96f4db6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.006102] env[63355]: DEBUG oslo_vmware.api [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1282.006102] env[63355]: value = "task-1350340" [ 1282.006102] env[63355]: _type = "Task" [ 1282.006102] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.013677] env[63355]: DEBUG oslo_vmware.api [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350340, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.090238] env[63355]: DEBUG oslo_concurrency.lockutils [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "e4c425fa-0b13-456c-b12b-c1ff18593536" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.090444] env[63355]: DEBUG oslo_concurrency.lockutils [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.517719] env[63355]: DEBUG oslo_vmware.api [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350340, 'name': ReconfigVM_Task, 'duration_secs': 0.125114} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.518020] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287781', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'name': 'volume-6f56474d-ef78-42b3-974e-5b060951acb4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b246e6a-9974-49e4-b213-7664c90e8420', 'attached_at': '', 'detached_at': '', 'volume_id': '6f56474d-ef78-42b3-974e-5b060951acb4', 'serial': '6f56474d-ef78-42b3-974e-5b060951acb4'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1282.594116] env[63355]: DEBUG nova.compute.utils [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1283.057572] env[63355]: DEBUG nova.objects.instance [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lazy-loading 'flavor' on Instance uuid 5b246e6a-9974-49e4-b213-7664c90e8420 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1283.097744] env[63355]: DEBUG oslo_concurrency.lockutils [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1284.064572] env[63355]: DEBUG oslo_concurrency.lockutils [None req-f77a0b0c-0007-41b6-ac3c-4bca66edfdd5 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.207s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1284.149450] env[63355]: DEBUG oslo_concurrency.lockutils [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "e4c425fa-0b13-456c-b12b-c1ff18593536" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1284.149698] env[63355]: DEBUG oslo_concurrency.lockutils [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1284.150329] env[63355]: INFO nova.compute.manager [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Attaching volume 0b48bb44-8c60-4c20-acdc-66e713b20d31 to /dev/sdb [ 1284.183512] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0001cc-500e-4e66-8d88-6d571f545d84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.190445] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db6f466-cd0d-4ed6-9c96-a9a22a9673ec {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.203410] env[63355]: DEBUG nova.virt.block_device [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Updating existing volume attachment record: 9e16ff5d-3289-49c6-ab28-d9089b50ce4c {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1285.107963] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "5b246e6a-9974-49e4-b213-7664c90e8420" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1285.108367] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1285.108678] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "5b246e6a-9974-49e4-b213-7664c90e8420-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1285.108736] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1285.108872] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1285.111078] env[63355]: INFO nova.compute.manager [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Terminating instance [ 1285.112822] env[63355]: DEBUG nova.compute.manager [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1285.113048] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1285.113846] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48567f0f-3423-4975-9724-030c62e65080 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.121471] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1285.121691] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d8e9aed-bbd9-450c-9fcd-779920057c8a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.127845] env[63355]: DEBUG oslo_vmware.api [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1285.127845] env[63355]: value = "task-1350342" [ 1285.127845] env[63355]: _type = "Task" [ 1285.127845] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.136669] env[63355]: DEBUG oslo_vmware.api [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350342, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.636939] env[63355]: DEBUG oslo_vmware.api [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350342, 'name': PowerOffVM_Task, 'duration_secs': 0.21279} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.637262] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1285.637436] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1285.637706] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f7c9bb49-0592-4ef1-87c1-b7af962e1fa4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.697565] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1285.697805] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1285.697956] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleting the datastore file [datastore1] 5b246e6a-9974-49e4-b213-7664c90e8420 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1285.698259] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cd85aff0-3e32-45f3-98ac-0ef7423d43ac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.705103] env[63355]: DEBUG oslo_vmware.api [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1285.705103] env[63355]: value = "task-1350344" [ 1285.705103] env[63355]: _type = "Task" [ 1285.705103] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.713864] env[63355]: DEBUG oslo_vmware.api [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350344, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.214794] env[63355]: DEBUG oslo_vmware.api [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350344, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126826} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.215180] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1286.215344] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1286.215539] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1286.215716] env[63355]: INFO nova.compute.manager [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1286.215954] env[63355]: DEBUG oslo.service.loopingcall [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1286.216166] env[63355]: DEBUG nova.compute.manager [-] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1286.216260] env[63355]: DEBUG nova.network.neutron [-] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1286.787877] env[63355]: DEBUG nova.compute.manager [req-907f80ff-dc73-4075-9d20-8ba67cf358b1 req-74d6c811-22e9-4d9c-b92c-e9264cd2424c service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Received event network-vif-deleted-313f0c1d-9da1-41ec-bba9-5ff65d52db63 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1286.788101] env[63355]: INFO nova.compute.manager [req-907f80ff-dc73-4075-9d20-8ba67cf358b1 req-74d6c811-22e9-4d9c-b92c-e9264cd2424c service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Neutron deleted interface 313f0c1d-9da1-41ec-bba9-5ff65d52db63; detaching it from the instance and deleting it from the info cache [ 1286.788252] env[63355]: DEBUG nova.network.neutron [req-907f80ff-dc73-4075-9d20-8ba67cf358b1 req-74d6c811-22e9-4d9c-b92c-e9264cd2424c service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1287.269247] env[63355]: DEBUG nova.network.neutron [-] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1287.291812] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c1b13b9-3a81-47d6-8799-881bc53f1aeb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.301762] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f80bb1-0391-4a39-837c-6b00c55bd0f1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.324887] env[63355]: DEBUG nova.compute.manager [req-907f80ff-dc73-4075-9d20-8ba67cf358b1 req-74d6c811-22e9-4d9c-b92c-e9264cd2424c service nova] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Detach interface failed, port_id=313f0c1d-9da1-41ec-bba9-5ff65d52db63, reason: Instance 5b246e6a-9974-49e4-b213-7664c90e8420 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1287.772146] env[63355]: INFO nova.compute.manager [-] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Took 1.56 seconds to deallocate network for instance. [ 1288.278640] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1288.279052] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1288.279216] env[63355]: DEBUG nova.objects.instance [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lazy-loading 'resources' on Instance uuid 5b246e6a-9974-49e4-b213-7664c90e8420 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1288.747388] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1288.747708] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287783', 'volume_id': '0b48bb44-8c60-4c20-acdc-66e713b20d31', 'name': 'volume-0b48bb44-8c60-4c20-acdc-66e713b20d31', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e4c425fa-0b13-456c-b12b-c1ff18593536', 'attached_at': '', 'detached_at': '', 'volume_id': '0b48bb44-8c60-4c20-acdc-66e713b20d31', 'serial': '0b48bb44-8c60-4c20-acdc-66e713b20d31'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1288.748636] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce1e5fda-84cc-4c14-af33-7717d4ec13c1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.765502] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f343cf75-ef89-4c07-aa7b-631d999c9166 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.792199] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] volume-0b48bb44-8c60-4c20-acdc-66e713b20d31/volume-0b48bb44-8c60-4c20-acdc-66e713b20d31.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1288.792642] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b11e2bb5-be01-4bd9-8ced-d47949844968 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.809738] env[63355]: DEBUG oslo_vmware.api [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1288.809738] env[63355]: value = "task-1350346" [ 1288.809738] env[63355]: _type = "Task" [ 1288.809738] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.819901] env[63355]: DEBUG oslo_vmware.api [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350346, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.845498] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd36cec-ff91-4982-a7d3-ea28db3c50f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.852042] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bed7c2-8974-4f28-8721-a9e192b1682d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.880365] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213033ca-798e-4f49-bc80-333230f9d481 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.886702] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54349431-7dec-4f5e-b1ab-e9acdd7c2a49 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.898868] env[63355]: DEBUG nova.compute.provider_tree [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1289.321758] env[63355]: DEBUG oslo_vmware.api [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350346, 'name': ReconfigVM_Task, 'duration_secs': 0.303222} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.322143] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Reconfigured VM instance instance-0000006d to attach disk [datastore2] volume-0b48bb44-8c60-4c20-acdc-66e713b20d31/volume-0b48bb44-8c60-4c20-acdc-66e713b20d31.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1289.326561] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b23965cf-82f8-4af0-86f3-8ae9c8481d9d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.340985] env[63355]: DEBUG oslo_vmware.api [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1289.340985] env[63355]: value = "task-1350347" [ 1289.340985] env[63355]: _type = "Task" [ 1289.340985] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.348145] env[63355]: DEBUG oslo_vmware.api [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350347, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.402522] env[63355]: DEBUG nova.scheduler.client.report [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1289.850549] env[63355]: DEBUG oslo_vmware.api [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350347, 'name': ReconfigVM_Task, 'duration_secs': 0.136466} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.850871] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287783', 'volume_id': '0b48bb44-8c60-4c20-acdc-66e713b20d31', 'name': 'volume-0b48bb44-8c60-4c20-acdc-66e713b20d31', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e4c425fa-0b13-456c-b12b-c1ff18593536', 'attached_at': '', 'detached_at': '', 'volume_id': '0b48bb44-8c60-4c20-acdc-66e713b20d31', 'serial': '0b48bb44-8c60-4c20-acdc-66e713b20d31'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1289.908014] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.629s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1289.927962] env[63355]: INFO nova.scheduler.client.report [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted allocations for instance 5b246e6a-9974-49e4-b213-7664c90e8420 [ 1290.437041] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d071c478-0a0e-41f3-8fe5-ddeb4a35c702 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "5b246e6a-9974-49e4-b213-7664c90e8420" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.328s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1290.891096] env[63355]: DEBUG nova.objects.instance [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lazy-loading 'flavor' on Instance uuid e4c425fa-0b13-456c-b12b-c1ff18593536 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1291.396344] env[63355]: DEBUG oslo_concurrency.lockutils [None req-925a19dc-5df5-4834-bbe8-8cb1b97c5afc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.246s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1291.555010] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "09e632ab-cd73-48cd-a4f3-308c247794c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1291.555293] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "09e632ab-cd73-48cd-a4f3-308c247794c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1291.599157] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "e4c425fa-0b13-456c-b12b-c1ff18593536" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1291.599384] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1292.057945] env[63355]: DEBUG nova.compute.manager [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1292.101760] env[63355]: INFO nova.compute.manager [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Detaching volume 0b48bb44-8c60-4c20-acdc-66e713b20d31 [ 1292.139541] env[63355]: INFO nova.virt.block_device [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Attempting to driver detach volume 0b48bb44-8c60-4c20-acdc-66e713b20d31 from mountpoint /dev/sdb [ 1292.139790] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1292.140058] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287783', 'volume_id': '0b48bb44-8c60-4c20-acdc-66e713b20d31', 'name': 'volume-0b48bb44-8c60-4c20-acdc-66e713b20d31', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e4c425fa-0b13-456c-b12b-c1ff18593536', 'attached_at': '', 'detached_at': '', 'volume_id': '0b48bb44-8c60-4c20-acdc-66e713b20d31', 'serial': '0b48bb44-8c60-4c20-acdc-66e713b20d31'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1292.140935] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2731ea4e-8e6e-4978-a094-9aca948543d4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.162969] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0021d46f-efe7-4554-94a5-8dd93256e5d9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.169177] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddb6aa1-8612-4180-8843-d7def7f314b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.188355] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b927b18-8441-4fe9-919e-4a7904f9ed3e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.203314] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] The volume has not been displaced from its original location: [datastore2] volume-0b48bb44-8c60-4c20-acdc-66e713b20d31/volume-0b48bb44-8c60-4c20-acdc-66e713b20d31.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1292.208384] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1292.208663] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0bcf3a1-c070-40fe-9ea8-f28ed4d3bfea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.225912] env[63355]: DEBUG oslo_vmware.api [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1292.225912] env[63355]: value = "task-1350348" [ 1292.225912] env[63355]: _type = "Task" [ 1292.225912] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.233328] env[63355]: DEBUG oslo_vmware.api [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350348, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.579029] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1292.579029] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1292.580409] env[63355]: INFO nova.compute.claims [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1292.735185] env[63355]: DEBUG oslo_vmware.api [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350348, 'name': ReconfigVM_Task, 'duration_secs': 0.199903} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.735479] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1292.740046] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-087aaf92-1e83-440f-9081-6c6ff3c00254 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.755113] env[63355]: DEBUG oslo_vmware.api [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1292.755113] env[63355]: value = "task-1350349" [ 1292.755113] env[63355]: _type = "Task" [ 1292.755113] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.763607] env[63355]: DEBUG oslo_vmware.api [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350349, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.264571] env[63355]: DEBUG oslo_vmware.api [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350349, 'name': ReconfigVM_Task, 'duration_secs': 0.132406} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.264838] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287783', 'volume_id': '0b48bb44-8c60-4c20-acdc-66e713b20d31', 'name': 'volume-0b48bb44-8c60-4c20-acdc-66e713b20d31', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e4c425fa-0b13-456c-b12b-c1ff18593536', 'attached_at': '', 'detached_at': '', 'volume_id': '0b48bb44-8c60-4c20-acdc-66e713b20d31', 'serial': '0b48bb44-8c60-4c20-acdc-66e713b20d31'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1293.633553] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71c28b0-f1de-4863-adca-f630df903598 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.641323] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d1c47b-a9f4-4f1b-8c2c-2c2a92f1455d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.670988] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3b29ab-43c8-4936-bcce-7ea42bfe82ce {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.677986] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae99ea4-9909-4f96-bd59-30c9a75b1647 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.690520] env[63355]: DEBUG nova.compute.provider_tree [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1293.803933] env[63355]: DEBUG nova.objects.instance [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lazy-loading 'flavor' on Instance uuid e4c425fa-0b13-456c-b12b-c1ff18593536 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1294.193620] env[63355]: DEBUG nova.scheduler.client.report [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1294.698022] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1294.698569] env[63355]: DEBUG nova.compute.manager [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1294.811453] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2d38c793-067a-4151-81bf-4fa64075a0bb tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.212s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.203398] env[63355]: DEBUG nova.compute.utils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1295.204781] env[63355]: DEBUG nova.compute.manager [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1295.204949] env[63355]: DEBUG nova.network.neutron [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1295.251925] env[63355]: DEBUG nova.policy [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a61bc098ea4e43ceab263ae9890da7c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff5b796e7bb0458dbd3b862dcae7bdd1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1295.500018] env[63355]: DEBUG nova.network.neutron [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Successfully created port: 7cec8c38-1693-4dde-b2b2-1165d43964e4 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1295.708328] env[63355]: DEBUG nova.compute.manager [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1295.833021] env[63355]: DEBUG oslo_concurrency.lockutils [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "e4c425fa-0b13-456c-b12b-c1ff18593536" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.833305] env[63355]: DEBUG oslo_concurrency.lockutils [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1295.833519] env[63355]: DEBUG oslo_concurrency.lockutils [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "e4c425fa-0b13-456c-b12b-c1ff18593536-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.833709] env[63355]: DEBUG oslo_concurrency.lockutils [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1295.833890] env[63355]: DEBUG oslo_concurrency.lockutils [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.835892] env[63355]: INFO nova.compute.manager [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Terminating instance [ 1295.838957] env[63355]: DEBUG nova.compute.manager [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1295.839177] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1295.840074] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a14815-2cb6-4af1-afed-bdfce78fb0da {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.847663] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1295.847897] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-080c5ff5-6812-481b-a348-64d8e743abe7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.853767] env[63355]: DEBUG oslo_vmware.api [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1295.853767] env[63355]: value = "task-1350350" [ 1295.853767] env[63355]: _type = "Task" [ 1295.853767] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.860819] env[63355]: DEBUG oslo_vmware.api [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350350, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.363551] env[63355]: DEBUG oslo_vmware.api [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350350, 'name': PowerOffVM_Task, 'duration_secs': 0.173829} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.363845] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1296.364007] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1296.364263] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4acf082-d6d6-422c-ba63-ac7bb3a1e2fb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.432110] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1296.432384] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1296.432578] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleting the datastore file [datastore2] e4c425fa-0b13-456c-b12b-c1ff18593536 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1296.432846] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ce1d882-adca-47d1-952b-7b2985e463c2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.439752] env[63355]: DEBUG oslo_vmware.api [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1296.439752] env[63355]: value = "task-1350352" [ 1296.439752] env[63355]: _type = "Task" [ 1296.439752] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.447170] env[63355]: DEBUG oslo_vmware.api [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350352, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.718246] env[63355]: DEBUG nova.compute.manager [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1296.743877] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1296.743877] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1296.744058] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1296.744247] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1296.744391] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1296.744539] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1296.744748] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1296.744910] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1296.745533] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1296.745533] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1296.745533] env[63355]: DEBUG nova.virt.hardware [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1296.746296] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326c7d69-864b-4210-866c-022261e99f69 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.753941] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365be35a-2ab1-4a61-a24e-ee5d15cf0372 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.861773] env[63355]: DEBUG nova.compute.manager [req-691f1047-37b1-464b-9083-1c80f316dc22 req-6d2446f8-2039-4eee-af54-a48a59627a1d service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Received event network-vif-plugged-7cec8c38-1693-4dde-b2b2-1165d43964e4 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1296.862014] env[63355]: DEBUG oslo_concurrency.lockutils [req-691f1047-37b1-464b-9083-1c80f316dc22 req-6d2446f8-2039-4eee-af54-a48a59627a1d service nova] Acquiring lock "09e632ab-cd73-48cd-a4f3-308c247794c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1296.862236] env[63355]: DEBUG oslo_concurrency.lockutils [req-691f1047-37b1-464b-9083-1c80f316dc22 req-6d2446f8-2039-4eee-af54-a48a59627a1d service nova] Lock "09e632ab-cd73-48cd-a4f3-308c247794c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1296.862406] env[63355]: DEBUG oslo_concurrency.lockutils [req-691f1047-37b1-464b-9083-1c80f316dc22 req-6d2446f8-2039-4eee-af54-a48a59627a1d service nova] Lock "09e632ab-cd73-48cd-a4f3-308c247794c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1296.862574] env[63355]: DEBUG nova.compute.manager [req-691f1047-37b1-464b-9083-1c80f316dc22 req-6d2446f8-2039-4eee-af54-a48a59627a1d service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] No waiting events found dispatching network-vif-plugged-7cec8c38-1693-4dde-b2b2-1165d43964e4 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1296.862738] env[63355]: WARNING nova.compute.manager [req-691f1047-37b1-464b-9083-1c80f316dc22 req-6d2446f8-2039-4eee-af54-a48a59627a1d service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Received unexpected event network-vif-plugged-7cec8c38-1693-4dde-b2b2-1165d43964e4 for instance with vm_state building and task_state spawning. [ 1296.944432] env[63355]: DEBUG nova.network.neutron [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Successfully updated port: 7cec8c38-1693-4dde-b2b2-1165d43964e4 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1296.951230] env[63355]: DEBUG oslo_vmware.api [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350352, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133715} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.951496] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1296.951683] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1296.951869] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1296.952062] env[63355]: INFO nova.compute.manager [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1296.952318] env[63355]: DEBUG oslo.service.loopingcall [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1296.952587] env[63355]: DEBUG nova.compute.manager [-] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1296.952587] env[63355]: DEBUG nova.network.neutron [-] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1297.447162] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "refresh_cache-09e632ab-cd73-48cd-a4f3-308c247794c8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1297.447331] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "refresh_cache-09e632ab-cd73-48cd-a4f3-308c247794c8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1297.447490] env[63355]: DEBUG nova.network.neutron [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1297.853466] env[63355]: DEBUG nova.network.neutron [-] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1297.978785] env[63355]: DEBUG nova.network.neutron [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1298.099594] env[63355]: DEBUG nova.network.neutron [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Updating instance_info_cache with network_info: [{"id": "7cec8c38-1693-4dde-b2b2-1165d43964e4", "address": "fa:16:3e:10:a9:73", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cec8c38-16", "ovs_interfaceid": "7cec8c38-1693-4dde-b2b2-1165d43964e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1298.358099] env[63355]: INFO nova.compute.manager [-] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Took 1.41 seconds to deallocate network for instance. [ 1298.602389] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "refresh_cache-09e632ab-cd73-48cd-a4f3-308c247794c8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1298.602748] env[63355]: DEBUG nova.compute.manager [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Instance network_info: |[{"id": "7cec8c38-1693-4dde-b2b2-1165d43964e4", "address": "fa:16:3e:10:a9:73", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cec8c38-16", "ovs_interfaceid": "7cec8c38-1693-4dde-b2b2-1165d43964e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1298.603224] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:a9:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7cec8c38-1693-4dde-b2b2-1165d43964e4', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1298.610769] env[63355]: DEBUG oslo.service.loopingcall [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1298.611012] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1298.611250] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-138f9a4c-5f09-440e-b3dd-20e88ddbc87e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.630659] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1298.630659] env[63355]: value = "task-1350353" [ 1298.630659] env[63355]: _type = "Task" [ 1298.630659] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.638114] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350353, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.864314] env[63355]: DEBUG oslo_concurrency.lockutils [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1298.864644] env[63355]: DEBUG oslo_concurrency.lockutils [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1298.864812] env[63355]: DEBUG nova.objects.instance [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lazy-loading 'resources' on Instance uuid e4c425fa-0b13-456c-b12b-c1ff18593536 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1298.889659] env[63355]: DEBUG nova.compute.manager [req-c9a31c0d-46f5-4045-8771-7fcf1c3665a1 req-048dca1f-8ca5-4179-b3f5-4c38165cd07b service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Received event network-changed-7cec8c38-1693-4dde-b2b2-1165d43964e4 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1298.889819] env[63355]: DEBUG nova.compute.manager [req-c9a31c0d-46f5-4045-8771-7fcf1c3665a1 req-048dca1f-8ca5-4179-b3f5-4c38165cd07b service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Refreshing instance network info cache due to event network-changed-7cec8c38-1693-4dde-b2b2-1165d43964e4. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1298.890124] env[63355]: DEBUG oslo_concurrency.lockutils [req-c9a31c0d-46f5-4045-8771-7fcf1c3665a1 req-048dca1f-8ca5-4179-b3f5-4c38165cd07b service nova] Acquiring lock "refresh_cache-09e632ab-cd73-48cd-a4f3-308c247794c8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1298.890228] env[63355]: DEBUG oslo_concurrency.lockutils [req-c9a31c0d-46f5-4045-8771-7fcf1c3665a1 req-048dca1f-8ca5-4179-b3f5-4c38165cd07b service nova] Acquired lock "refresh_cache-09e632ab-cd73-48cd-a4f3-308c247794c8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1298.890390] env[63355]: DEBUG nova.network.neutron [req-c9a31c0d-46f5-4045-8771-7fcf1c3665a1 req-048dca1f-8ca5-4179-b3f5-4c38165cd07b service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Refreshing network info cache for port 7cec8c38-1693-4dde-b2b2-1165d43964e4 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1299.140552] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350353, 'name': CreateVM_Task, 'duration_secs': 0.321489} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1299.140717] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1299.141415] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1299.141589] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1299.141921] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1299.142191] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6cee38c-df35-4a8e-acbf-079a6e54fe89 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.146274] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1299.146274] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522a0ece-f30f-8293-271a-a180ac894a6b" [ 1299.146274] env[63355]: _type = "Task" [ 1299.146274] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1299.153518] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522a0ece-f30f-8293-271a-a180ac894a6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.419170] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf709f55-e76b-4914-98e6-281901374b4e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.428293] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db416af-880d-4f26-8f16-883cbf262674 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.458158] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7994d9-e8de-47ea-bdfe-343aba8a76b4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.466459] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf80258-06da-4926-b520-cbd12a3de959 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.479042] env[63355]: DEBUG nova.compute.provider_tree [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1299.585669] env[63355]: DEBUG nova.network.neutron [req-c9a31c0d-46f5-4045-8771-7fcf1c3665a1 req-048dca1f-8ca5-4179-b3f5-4c38165cd07b service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Updated VIF entry in instance network info cache for port 7cec8c38-1693-4dde-b2b2-1165d43964e4. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1299.586044] env[63355]: DEBUG nova.network.neutron [req-c9a31c0d-46f5-4045-8771-7fcf1c3665a1 req-048dca1f-8ca5-4179-b3f5-4c38165cd07b service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Updating instance_info_cache with network_info: [{"id": "7cec8c38-1693-4dde-b2b2-1165d43964e4", "address": "fa:16:3e:10:a9:73", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cec8c38-16", "ovs_interfaceid": "7cec8c38-1693-4dde-b2b2-1165d43964e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1299.656019] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522a0ece-f30f-8293-271a-a180ac894a6b, 'name': SearchDatastore_Task, 'duration_secs': 0.00881} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1299.656366] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1299.656599] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1299.656831] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1299.656980] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1299.657193] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1299.657446] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-417dcbdf-a12b-498d-97b5-f5c987c89bba {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.664960] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1299.665155] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1299.665802] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4164b5b-e0c6-4a5b-897d-b691e9845851 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.670257] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1299.670257] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529392b2-dc51-1bc4-dcee-22dbccabd0d3" [ 1299.670257] env[63355]: _type = "Task" [ 1299.670257] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1299.677120] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529392b2-dc51-1bc4-dcee-22dbccabd0d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.984026] env[63355]: DEBUG nova.scheduler.client.report [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1300.088754] env[63355]: DEBUG oslo_concurrency.lockutils [req-c9a31c0d-46f5-4045-8771-7fcf1c3665a1 req-048dca1f-8ca5-4179-b3f5-4c38165cd07b service nova] Releasing lock "refresh_cache-09e632ab-cd73-48cd-a4f3-308c247794c8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1300.089110] env[63355]: DEBUG nova.compute.manager [req-c9a31c0d-46f5-4045-8771-7fcf1c3665a1 req-048dca1f-8ca5-4179-b3f5-4c38165cd07b service nova] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Received event network-vif-deleted-bc7a8f76-e220-4474-8089-83364388f489 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1300.180319] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]529392b2-dc51-1bc4-dcee-22dbccabd0d3, 'name': SearchDatastore_Task, 'duration_secs': 0.008299} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.181178] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e36f86b-9bd3-4986-89f2-1186a1d573fa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.186017] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1300.186017] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527755b1-9d0b-3d23-a232-c182ca9ffe39" [ 1300.186017] env[63355]: _type = "Task" [ 1300.186017] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.192977] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527755b1-9d0b-3d23-a232-c182ca9ffe39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.488666] env[63355]: DEBUG oslo_concurrency.lockutils [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.624s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1300.509122] env[63355]: INFO nova.scheduler.client.report [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleted allocations for instance e4c425fa-0b13-456c-b12b-c1ff18593536 [ 1300.696459] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]527755b1-9d0b-3d23-a232-c182ca9ffe39, 'name': SearchDatastore_Task, 'duration_secs': 0.01004} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.696753] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1300.697035] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 09e632ab-cd73-48cd-a4f3-308c247794c8/09e632ab-cd73-48cd-a4f3-308c247794c8.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1300.697323] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa0ecdf1-887c-49cf-88be-b0322ab737e6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.703701] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1300.703701] env[63355]: value = "task-1350354" [ 1300.703701] env[63355]: _type = "Task" [ 1300.703701] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.711215] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350354, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.017276] env[63355]: DEBUG oslo_concurrency.lockutils [None req-51b780db-8279-48eb-8d10-9e0f2aba59e3 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "e4c425fa-0b13-456c-b12b-c1ff18593536" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.184s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1301.215353] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350354, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507569} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.215653] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore1] 09e632ab-cd73-48cd-a4f3-308c247794c8/09e632ab-cd73-48cd-a4f3-308c247794c8.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1301.215874] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1301.216164] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-05437a00-d9ac-41b7-ab3d-bca0048a9e68 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.222747] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1301.222747] env[63355]: value = "task-1350355" [ 1301.222747] env[63355]: _type = "Task" [ 1301.222747] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.230374] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350355, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.733212] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350355, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.156111} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.733633] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1301.734978] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b4f1998-a1fe-49d4-a0ab-b38460c5fe53 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.757736] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 09e632ab-cd73-48cd-a4f3-308c247794c8/09e632ab-cd73-48cd-a4f3-308c247794c8.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1301.758134] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b15e095b-9127-417e-94d5-981e2aafde9b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.778326] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1301.778326] env[63355]: value = "task-1350356" [ 1301.778326] env[63355]: _type = "Task" [ 1301.778326] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.785899] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350356, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.288059] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350356, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.791589] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "9f7c978f-e993-4fa9-9402-489545a10020" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1302.791815] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1302.793224] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350356, 'name': ReconfigVM_Task, 'duration_secs': 0.736632} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.793781] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 09e632ab-cd73-48cd-a4f3-308c247794c8/09e632ab-cd73-48cd-a4f3-308c247794c8.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1302.794340] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3decfb6f-5de8-4316-aa78-2cb586bea2a9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.801063] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1302.801063] env[63355]: value = "task-1350357" [ 1302.801063] env[63355]: _type = "Task" [ 1302.801063] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.810588] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350357, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.294222] env[63355]: DEBUG nova.compute.manager [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1303.310504] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350357, 'name': Rename_Task, 'duration_secs': 0.150057} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.310605] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1303.310830] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d139ff98-68cf-45ba-922a-75e1b601e5ac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.317025] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1303.317025] env[63355]: value = "task-1350358" [ 1303.317025] env[63355]: _type = "Task" [ 1303.317025] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.323950] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350358, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.813918] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1303.814202] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1303.815693] env[63355]: INFO nova.compute.claims [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1303.827363] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350358, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.332092] env[63355]: DEBUG oslo_vmware.api [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350358, 'name': PowerOnVM_Task, 'duration_secs': 0.62854} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.333047] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1304.333271] env[63355]: INFO nova.compute.manager [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Took 7.61 seconds to spawn the instance on the hypervisor. [ 1304.333453] env[63355]: DEBUG nova.compute.manager [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1304.334225] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edc2dff-6473-4d4b-a71b-c2948ef0c238 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.849046] env[63355]: DEBUG nova.scheduler.client.report [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Refreshing inventories for resource provider 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1304.853183] env[63355]: INFO nova.compute.manager [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Took 12.29 seconds to build instance. [ 1304.862643] env[63355]: DEBUG nova.scheduler.client.report [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Updating ProviderTree inventory for provider 47abb610-db7e-4770-911d-187dd075ef8b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1304.862873] env[63355]: DEBUG nova.compute.provider_tree [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Updating inventory in ProviderTree for provider 47abb610-db7e-4770-911d-187dd075ef8b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1304.873363] env[63355]: DEBUG nova.scheduler.client.report [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Refreshing aggregate associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, aggregates: None {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1304.890041] env[63355]: DEBUG nova.scheduler.client.report [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Refreshing trait associations for resource provider 47abb610-db7e-4770-911d-187dd075ef8b, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=63355) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1304.933102] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2950223c-d1c9-47da-9b2e-7026265f50b2 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.940808] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83512ade-deda-4d6c-b546-d25a3ba02109 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.970245] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08323bd6-1bbd-42a7-a483-88d80484ba2e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.977249] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bcc2a4-feca-4381-9cad-1a8228044dfb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.989860] env[63355]: DEBUG nova.compute.provider_tree [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1305.355601] env[63355]: DEBUG oslo_concurrency.lockutils [None req-7801ae09-5e6a-4806-bc30-35190cbed298 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "09e632ab-cd73-48cd-a4f3-308c247794c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.800s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1305.484197] env[63355]: DEBUG nova.compute.manager [req-ce036af3-aeb8-4adc-b28b-0b0b79691e7f req-7467bcde-9b49-4463-9374-725c0da684dc service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Received event network-changed-7cec8c38-1693-4dde-b2b2-1165d43964e4 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1305.484434] env[63355]: DEBUG nova.compute.manager [req-ce036af3-aeb8-4adc-b28b-0b0b79691e7f req-7467bcde-9b49-4463-9374-725c0da684dc service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Refreshing instance network info cache due to event network-changed-7cec8c38-1693-4dde-b2b2-1165d43964e4. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1305.484655] env[63355]: DEBUG oslo_concurrency.lockutils [req-ce036af3-aeb8-4adc-b28b-0b0b79691e7f req-7467bcde-9b49-4463-9374-725c0da684dc service nova] Acquiring lock "refresh_cache-09e632ab-cd73-48cd-a4f3-308c247794c8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1305.484803] env[63355]: DEBUG oslo_concurrency.lockutils [req-ce036af3-aeb8-4adc-b28b-0b0b79691e7f req-7467bcde-9b49-4463-9374-725c0da684dc service nova] Acquired lock "refresh_cache-09e632ab-cd73-48cd-a4f3-308c247794c8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1305.484981] env[63355]: DEBUG nova.network.neutron [req-ce036af3-aeb8-4adc-b28b-0b0b79691e7f req-7467bcde-9b49-4463-9374-725c0da684dc service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Refreshing network info cache for port 7cec8c38-1693-4dde-b2b2-1165d43964e4 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1305.492989] env[63355]: DEBUG nova.scheduler.client.report [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1305.997149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.183s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1305.997667] env[63355]: DEBUG nova.compute.manager [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1306.180576] env[63355]: DEBUG nova.network.neutron [req-ce036af3-aeb8-4adc-b28b-0b0b79691e7f req-7467bcde-9b49-4463-9374-725c0da684dc service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Updated VIF entry in instance network info cache for port 7cec8c38-1693-4dde-b2b2-1165d43964e4. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1306.180915] env[63355]: DEBUG nova.network.neutron [req-ce036af3-aeb8-4adc-b28b-0b0b79691e7f req-7467bcde-9b49-4463-9374-725c0da684dc service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Updating instance_info_cache with network_info: [{"id": "7cec8c38-1693-4dde-b2b2-1165d43964e4", "address": "fa:16:3e:10:a9:73", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cec8c38-16", "ovs_interfaceid": "7cec8c38-1693-4dde-b2b2-1165d43964e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1306.503092] env[63355]: DEBUG nova.compute.utils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1306.504500] env[63355]: DEBUG nova.compute.manager [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1306.504666] env[63355]: DEBUG nova.network.neutron [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1306.549432] env[63355]: DEBUG nova.policy [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7302e68a14984b08aef9c3af0d2a12ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '876995f179ed46b397822fa1be08ea29', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1306.683756] env[63355]: DEBUG oslo_concurrency.lockutils [req-ce036af3-aeb8-4adc-b28b-0b0b79691e7f req-7467bcde-9b49-4463-9374-725c0da684dc service nova] Releasing lock "refresh_cache-09e632ab-cd73-48cd-a4f3-308c247794c8" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1306.811974] env[63355]: DEBUG nova.network.neutron [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Successfully created port: 70f5b776-10c7-4fa8-be9e-a318da6adf88 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1307.007284] env[63355]: DEBUG nova.compute.manager [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1308.016615] env[63355]: DEBUG nova.compute.manager [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1308.042073] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T02:22:43Z,direct_url=,disk_format='vmdk',id=84ca0bb3-9916-4cea-9399-75af77b0558e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfa2720fe39d48deb0cad174a7e5f3ea',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T02:22:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1308.042362] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1308.042534] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1308.042721] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1308.042870] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1308.043032] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1308.043254] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1308.043416] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1308.043585] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1308.043751] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1308.043924] env[63355]: DEBUG nova.virt.hardware [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1308.044795] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87301f88-8bfe-4242-87c6-3a26d570aeaa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.052453] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f32f658-0a9c-4c25-874e-2bd6855474e8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.166019] env[63355]: DEBUG nova.compute.manager [req-3d676c19-a693-4692-9e3a-8ee33e497ba5 req-d4b39aee-532b-4dc0-bc50-67417db4cd2f service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Received event network-vif-plugged-70f5b776-10c7-4fa8-be9e-a318da6adf88 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1308.166288] env[63355]: DEBUG oslo_concurrency.lockutils [req-3d676c19-a693-4692-9e3a-8ee33e497ba5 req-d4b39aee-532b-4dc0-bc50-67417db4cd2f service nova] Acquiring lock "9f7c978f-e993-4fa9-9402-489545a10020-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1308.166449] env[63355]: DEBUG oslo_concurrency.lockutils [req-3d676c19-a693-4692-9e3a-8ee33e497ba5 req-d4b39aee-532b-4dc0-bc50-67417db4cd2f service nova] Lock "9f7c978f-e993-4fa9-9402-489545a10020-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1308.166620] env[63355]: DEBUG oslo_concurrency.lockutils [req-3d676c19-a693-4692-9e3a-8ee33e497ba5 req-d4b39aee-532b-4dc0-bc50-67417db4cd2f service nova] Lock "9f7c978f-e993-4fa9-9402-489545a10020-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1308.166787] env[63355]: DEBUG nova.compute.manager [req-3d676c19-a693-4692-9e3a-8ee33e497ba5 req-d4b39aee-532b-4dc0-bc50-67417db4cd2f service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] No waiting events found dispatching network-vif-plugged-70f5b776-10c7-4fa8-be9e-a318da6adf88 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1308.166949] env[63355]: WARNING nova.compute.manager [req-3d676c19-a693-4692-9e3a-8ee33e497ba5 req-d4b39aee-532b-4dc0-bc50-67417db4cd2f service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Received unexpected event network-vif-plugged-70f5b776-10c7-4fa8-be9e-a318da6adf88 for instance with vm_state building and task_state spawning. [ 1308.720988] env[63355]: DEBUG nova.network.neutron [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Successfully updated port: 70f5b776-10c7-4fa8-be9e-a318da6adf88 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1308.742312] env[63355]: DEBUG nova.compute.manager [req-bf37e216-db32-4fc4-8af7-574fb962cbbb req-94866504-bd0b-49fc-9ba4-a8590bdcaf51 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Received event network-changed-70f5b776-10c7-4fa8-be9e-a318da6adf88 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1308.742633] env[63355]: DEBUG nova.compute.manager [req-bf37e216-db32-4fc4-8af7-574fb962cbbb req-94866504-bd0b-49fc-9ba4-a8590bdcaf51 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Refreshing instance network info cache due to event network-changed-70f5b776-10c7-4fa8-be9e-a318da6adf88. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1308.742754] env[63355]: DEBUG oslo_concurrency.lockutils [req-bf37e216-db32-4fc4-8af7-574fb962cbbb req-94866504-bd0b-49fc-9ba4-a8590bdcaf51 service nova] Acquiring lock "refresh_cache-9f7c978f-e993-4fa9-9402-489545a10020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1308.742882] env[63355]: DEBUG oslo_concurrency.lockutils [req-bf37e216-db32-4fc4-8af7-574fb962cbbb req-94866504-bd0b-49fc-9ba4-a8590bdcaf51 service nova] Acquired lock "refresh_cache-9f7c978f-e993-4fa9-9402-489545a10020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1308.743833] env[63355]: DEBUG nova.network.neutron [req-bf37e216-db32-4fc4-8af7-574fb962cbbb req-94866504-bd0b-49fc-9ba4-a8590bdcaf51 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Refreshing network info cache for port 70f5b776-10c7-4fa8-be9e-a318da6adf88 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1309.222709] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "refresh_cache-9f7c978f-e993-4fa9-9402-489545a10020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1309.278355] env[63355]: DEBUG nova.network.neutron [req-bf37e216-db32-4fc4-8af7-574fb962cbbb req-94866504-bd0b-49fc-9ba4-a8590bdcaf51 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1309.364724] env[63355]: DEBUG nova.network.neutron [req-bf37e216-db32-4fc4-8af7-574fb962cbbb req-94866504-bd0b-49fc-9ba4-a8590bdcaf51 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1309.867923] env[63355]: DEBUG oslo_concurrency.lockutils [req-bf37e216-db32-4fc4-8af7-574fb962cbbb req-94866504-bd0b-49fc-9ba4-a8590bdcaf51 service nova] Releasing lock "refresh_cache-9f7c978f-e993-4fa9-9402-489545a10020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1309.868370] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "refresh_cache-9f7c978f-e993-4fa9-9402-489545a10020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1309.868532] env[63355]: DEBUG nova.network.neutron [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1310.133259] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1310.400585] env[63355]: DEBUG nova.network.neutron [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1310.533934] env[63355]: DEBUG nova.network.neutron [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Updating instance_info_cache with network_info: [{"id": "70f5b776-10c7-4fa8-be9e-a318da6adf88", "address": "fa:16:3e:11:f6:5e", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f5b776-10", "ovs_interfaceid": "70f5b776-10c7-4fa8-be9e-a318da6adf88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1310.636261] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1310.636506] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1310.636674] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1310.636835] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1310.637762] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f6d74f-b784-49b7-9398-df7160520fde {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.646603] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d55e9a-c6c6-490b-abec-a80790c20fe4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.660532] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62932a5-7ab6-444c-a8a8-1086a61a4055 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.666942] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375e8b47-1477-44c4-ada2-7fa13984ca3b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.696768] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181065MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1310.696981] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1310.697163] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1311.036656] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "refresh_cache-9f7c978f-e993-4fa9-9402-489545a10020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1311.036935] env[63355]: DEBUG nova.compute.manager [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Instance network_info: |[{"id": "70f5b776-10c7-4fa8-be9e-a318da6adf88", "address": "fa:16:3e:11:f6:5e", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f5b776-10", "ovs_interfaceid": "70f5b776-10c7-4fa8-be9e-a318da6adf88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1311.037410] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:f6:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '70f5b776-10c7-4fa8-be9e-a318da6adf88', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1311.045398] env[63355]: DEBUG oslo.service.loopingcall [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1311.045684] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1311.045921] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da3cb3c1-dca1-40da-a67e-aed8ef44df79 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.064610] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1311.064610] env[63355]: value = "task-1350361" [ 1311.064610] env[63355]: _type = "Task" [ 1311.064610] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.071761] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350361, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.574279] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350361, 'name': CreateVM_Task, 'duration_secs': 0.323397} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.574545] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1311.575131] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1311.575302] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1311.575627] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1311.575873] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d4a5401-a71a-48e9-bc0b-aa218adfd971 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.580776] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1311.580776] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522b72d4-e3e7-3e4d-fa8f-4d229a9c2ff6" [ 1311.580776] env[63355]: _type = "Task" [ 1311.580776] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.588037] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522b72d4-e3e7-3e4d-fa8f-4d229a9c2ff6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.721565] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8db23054-4556-496a-a4b1-a24c71c30f50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1311.721720] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 09e632ab-cd73-48cd-a4f3-308c247794c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1311.721842] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 9f7c978f-e993-4fa9-9402-489545a10020 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1311.722026] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1311.722176] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1311.764395] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d87db49-489a-4adb-a6c8-5662d3a7cef4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.771502] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fb0742-c692-4d2b-b814-9c3dd913051f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.799523] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f199361a-4e89-44bc-8160-fa50a8cc12cc {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.805770] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-079e312e-bdae-4bbb-b36f-84ca0331aafd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.819352] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1312.090803] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522b72d4-e3e7-3e4d-fa8f-4d229a9c2ff6, 'name': SearchDatastore_Task, 'duration_secs': 0.017132} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.091072] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1312.091345] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Processing image 84ca0bb3-9916-4cea-9399-75af77b0558e {{(pid=63355) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1312.091578] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1312.091727] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1312.091903] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1312.092164] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d33955f-d8ea-4a04-ba33-356107ce5cd7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.099809] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63355) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1312.099981] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63355) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1312.100659] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8509cec-3549-4d07-a988-c574bb8f446b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.105150] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1312.105150] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ad142a-3f86-e4db-4626-470dd5e0b943" [ 1312.105150] env[63355]: _type = "Task" [ 1312.105150] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.111906] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ad142a-3f86-e4db-4626-470dd5e0b943, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.322225] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1312.614920] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]52ad142a-3f86-e4db-4626-470dd5e0b943, 'name': SearchDatastore_Task, 'duration_secs': 0.007248} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.615698] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f038d1d6-d6fd-49c1-987f-e58c62d9cc3b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.620429] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1312.620429] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522b1d91-4522-8775-20c2-a55834060b07" [ 1312.620429] env[63355]: _type = "Task" [ 1312.620429] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.627399] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522b1d91-4522-8775-20c2-a55834060b07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.826498] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1312.826716] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.130s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1313.130437] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]522b1d91-4522-8775-20c2-a55834060b07, 'name': SearchDatastore_Task, 'duration_secs': 0.028413} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.130706] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1313.130964] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 9f7c978f-e993-4fa9-9402-489545a10020/9f7c978f-e993-4fa9-9402-489545a10020.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1313.131264] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9f249e1-1e3e-4bff-8efb-aad41849a21a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.137832] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1313.137832] env[63355]: value = "task-1350363" [ 1313.137832] env[63355]: _type = "Task" [ 1313.137832] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.144798] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350363, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.646406] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350363, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438987} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.646693] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/84ca0bb3-9916-4cea-9399-75af77b0558e/84ca0bb3-9916-4cea-9399-75af77b0558e.vmdk to [datastore2] 9f7c978f-e993-4fa9-9402-489545a10020/9f7c978f-e993-4fa9-9402-489545a10020.vmdk {{(pid=63355) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1313.646873] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Extending root virtual disk to 1048576 {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1313.647136] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3379eca5-23ff-4253-96b6-ad622d0566de {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.653692] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1313.653692] env[63355]: value = "task-1350365" [ 1313.653692] env[63355]: _type = "Task" [ 1313.653692] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.663015] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.164311] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059534} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.164311] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Extended root virtual disk {{(pid=63355) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1314.164621] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4b3b04-2563-40a3-8329-7ae3a3655577 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.185290] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 9f7c978f-e993-4fa9-9402-489545a10020/9f7c978f-e993-4fa9-9402-489545a10020.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1314.185544] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6f3327f-0cdf-4e3f-8bae-2749d2e87ee8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.203977] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1314.203977] env[63355]: value = "task-1350366" [ 1314.203977] env[63355]: _type = "Task" [ 1314.203977] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.211058] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350366, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.716460] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350366, 'name': ReconfigVM_Task, 'duration_secs': 0.269022} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.716752] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 9f7c978f-e993-4fa9-9402-489545a10020/9f7c978f-e993-4fa9-9402-489545a10020.vmdk or device None with type sparse {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1314.717358] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3929743-42f4-4026-9ed0-7733caca62f3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.722919] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1314.722919] env[63355]: value = "task-1350367" [ 1314.722919] env[63355]: _type = "Task" [ 1314.722919] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.730429] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350367, 'name': Rename_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.233257] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350367, 'name': Rename_Task, 'duration_secs': 0.132714} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.233546] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1315.233787] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec2ce961-a930-4897-ad60-2e46a8c51f07 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.239346] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1315.239346] env[63355]: value = "task-1350368" [ 1315.239346] env[63355]: _type = "Task" [ 1315.239346] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.246323] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350368, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.747981] env[63355]: DEBUG oslo_vmware.api [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350368, 'name': PowerOnVM_Task, 'duration_secs': 0.436486} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.748308] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1315.748561] env[63355]: INFO nova.compute.manager [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Took 7.73 seconds to spawn the instance on the hypervisor. [ 1315.748789] env[63355]: DEBUG nova.compute.manager [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1315.749535] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fac3c06-3f3f-4413-8124-6a6d808873af {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.267284] env[63355]: INFO nova.compute.manager [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Took 12.47 seconds to build instance. [ 1316.769220] env[63355]: DEBUG oslo_concurrency.lockutils [None req-eb55da10-ff41-46f6-ac45-4eac0c27d547 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.977s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1317.027264] env[63355]: DEBUG nova.compute.manager [req-534e0b4a-c0b9-41c7-a451-6babe3ee6456 req-e41095d9-c1a1-49a7-863d-5a12f83653e7 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Received event network-changed-70f5b776-10c7-4fa8-be9e-a318da6adf88 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1317.027264] env[63355]: DEBUG nova.compute.manager [req-534e0b4a-c0b9-41c7-a451-6babe3ee6456 req-e41095d9-c1a1-49a7-863d-5a12f83653e7 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Refreshing instance network info cache due to event network-changed-70f5b776-10c7-4fa8-be9e-a318da6adf88. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1317.027558] env[63355]: DEBUG oslo_concurrency.lockutils [req-534e0b4a-c0b9-41c7-a451-6babe3ee6456 req-e41095d9-c1a1-49a7-863d-5a12f83653e7 service nova] Acquiring lock "refresh_cache-9f7c978f-e993-4fa9-9402-489545a10020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1317.027558] env[63355]: DEBUG oslo_concurrency.lockutils [req-534e0b4a-c0b9-41c7-a451-6babe3ee6456 req-e41095d9-c1a1-49a7-863d-5a12f83653e7 service nova] Acquired lock "refresh_cache-9f7c978f-e993-4fa9-9402-489545a10020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1317.027720] env[63355]: DEBUG nova.network.neutron [req-534e0b4a-c0b9-41c7-a451-6babe3ee6456 req-e41095d9-c1a1-49a7-863d-5a12f83653e7 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Refreshing network info cache for port 70f5b776-10c7-4fa8-be9e-a318da6adf88 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1317.729606] env[63355]: DEBUG nova.network.neutron [req-534e0b4a-c0b9-41c7-a451-6babe3ee6456 req-e41095d9-c1a1-49a7-863d-5a12f83653e7 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Updated VIF entry in instance network info cache for port 70f5b776-10c7-4fa8-be9e-a318da6adf88. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1317.730127] env[63355]: DEBUG nova.network.neutron [req-534e0b4a-c0b9-41c7-a451-6babe3ee6456 req-e41095d9-c1a1-49a7-863d-5a12f83653e7 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Updating instance_info_cache with network_info: [{"id": "70f5b776-10c7-4fa8-be9e-a318da6adf88", "address": "fa:16:3e:11:f6:5e", "network": {"id": "6945b028-33a5-4d1d-910e-a74bd0ad76d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-526117092-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "876995f179ed46b397822fa1be08ea29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f5b776-10", "ovs_interfaceid": "70f5b776-10c7-4fa8-be9e-a318da6adf88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1317.822619] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1317.822961] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1317.822961] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1318.233293] env[63355]: DEBUG oslo_concurrency.lockutils [req-534e0b4a-c0b9-41c7-a451-6babe3ee6456 req-e41095d9-c1a1-49a7-863d-5a12f83653e7 service nova] Releasing lock "refresh_cache-9f7c978f-e993-4fa9-9402-489545a10020" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1319.835115] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Didn't find any instances for network info cache update. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1319.835389] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.835545] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.835641] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.835787] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.835925] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.836080] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.836213] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1326.996836] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "c2629183-3353-4bc6-8bf6-8fc72a874620" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.997129] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1327.499984] env[63355]: DEBUG nova.compute.manager [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Starting instance... {{(pid=63355) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1328.019314] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1328.019574] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1328.021117] env[63355]: INFO nova.compute.claims [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1329.084263] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6893f6f-0e8f-4133-bf77-405058d09a8d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.092246] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e568e4-41e9-45e1-9cbb-fd4b474ddb13 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.120693] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66962cb-3711-47da-9c0e-4ea7d96354f3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.127188] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3f2cf9-a59d-45ce-b78a-7c26279ca35a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.140643] env[63355]: DEBUG nova.compute.provider_tree [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1329.644338] env[63355]: DEBUG nova.scheduler.client.report [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1330.149999] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.130s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.150514] env[63355]: DEBUG nova.compute.manager [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Start building networks asynchronously for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1330.655303] env[63355]: DEBUG nova.compute.utils [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1330.656716] env[63355]: DEBUG nova.compute.manager [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Allocating IP information in the background. {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1330.656887] env[63355]: DEBUG nova.network.neutron [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] allocate_for_instance() {{(pid=63355) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1330.707715] env[63355]: DEBUG nova.policy [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a61bc098ea4e43ceab263ae9890da7c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff5b796e7bb0458dbd3b862dcae7bdd1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63355) authorize /opt/stack/nova/nova/policy.py:201}} [ 1330.955794] env[63355]: DEBUG nova.network.neutron [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Successfully created port: 1f3d442f-7173-41c7-b933-3cddb214c854 {{(pid=63355) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1331.160906] env[63355]: DEBUG nova.compute.manager [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Start building block device mappings for instance. {{(pid=63355) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1331.666243] env[63355]: INFO nova.virt.block_device [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Booting with volume f9514f73-1171-47d3-92bf-d2c1660b77d0 at /dev/sda [ 1331.702019] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9d2774e-2e5b-401f-9e07-fe4a26a68e36 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.712727] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60783361-71b4-4eb5-8371-a6fb60c5431b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.746140] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0afe4b91-105a-4248-83dd-089edd39e704 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.756206] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e396dd-aaa7-4773-9265-85ef5cdce670 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.781380] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d490fdea-463c-44fc-9be2-182dcef6123d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.787334] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bef5b4-3234-4a7f-893b-dec42e725df8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.800655] env[63355]: DEBUG nova.virt.block_device [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating existing volume attachment record: 6e927790-be9e-4e93-9ebc-d64583996bd7 {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1332.314125] env[63355]: DEBUG nova.compute.manager [req-de9978b5-41b7-4722-8de2-aa235c412ede req-9c214fe1-3d41-499d-91a0-505b48171ae3 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Received event network-vif-plugged-1f3d442f-7173-41c7-b933-3cddb214c854 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1332.314401] env[63355]: DEBUG oslo_concurrency.lockutils [req-de9978b5-41b7-4722-8de2-aa235c412ede req-9c214fe1-3d41-499d-91a0-505b48171ae3 service nova] Acquiring lock "c2629183-3353-4bc6-8bf6-8fc72a874620-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1332.314854] env[63355]: DEBUG oslo_concurrency.lockutils [req-de9978b5-41b7-4722-8de2-aa235c412ede req-9c214fe1-3d41-499d-91a0-505b48171ae3 service nova] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1332.314854] env[63355]: DEBUG oslo_concurrency.lockutils [req-de9978b5-41b7-4722-8de2-aa235c412ede req-9c214fe1-3d41-499d-91a0-505b48171ae3 service nova] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1332.315157] env[63355]: DEBUG nova.compute.manager [req-de9978b5-41b7-4722-8de2-aa235c412ede req-9c214fe1-3d41-499d-91a0-505b48171ae3 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] No waiting events found dispatching network-vif-plugged-1f3d442f-7173-41c7-b933-3cddb214c854 {{(pid=63355) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1332.315157] env[63355]: WARNING nova.compute.manager [req-de9978b5-41b7-4722-8de2-aa235c412ede req-9c214fe1-3d41-499d-91a0-505b48171ae3 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Received unexpected event network-vif-plugged-1f3d442f-7173-41c7-b933-3cddb214c854 for instance with vm_state building and task_state block_device_mapping. [ 1332.396441] env[63355]: DEBUG nova.network.neutron [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Successfully updated port: 1f3d442f-7173-41c7-b933-3cddb214c854 {{(pid=63355) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1332.899218] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1332.899218] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1332.899392] env[63355]: DEBUG nova.network.neutron [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1333.431194] env[63355]: DEBUG nova.network.neutron [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Instance cache missing network info. {{(pid=63355) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1333.549663] env[63355]: DEBUG nova.network.neutron [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance_info_cache with network_info: [{"id": "1f3d442f-7173-41c7-b933-3cddb214c854", "address": "fa:16:3e:e8:8f:00", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f3d442f-71", "ovs_interfaceid": "1f3d442f-7173-41c7-b933-3cddb214c854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1333.881528] env[63355]: DEBUG nova.compute.manager [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Start spawning the instance on the hypervisor. {{(pid=63355) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1333.882085] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1333.882313] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1333.882474] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1333.882692] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1333.882801] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1333.882949] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1333.883171] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1333.883333] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1333.883504] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1333.883668] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1333.883841] env[63355]: DEBUG nova.virt.hardware [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1333.884712] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d59a67-3bb3-4f1a-b0b4-0bc6c715b2f6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.893215] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3bf7d1-6afb-4dad-8b71-2b223fa83ceb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.052694] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1334.053019] env[63355]: DEBUG nova.compute.manager [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Instance network_info: |[{"id": "1f3d442f-7173-41c7-b933-3cddb214c854", "address": "fa:16:3e:e8:8f:00", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f3d442f-71", "ovs_interfaceid": "1f3d442f-7173-41c7-b933-3cddb214c854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63355) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1334.053466] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:8f:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1f3d442f-7173-41c7-b933-3cddb214c854', 'vif_model': 'vmxnet3'}] {{(pid=63355) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1334.061158] env[63355]: DEBUG oslo.service.loopingcall [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1334.061375] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Creating VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1334.061594] env[63355]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a8df5824-d54b-45e8-a623-7ee0d060da8d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.083939] env[63355]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1334.083939] env[63355]: value = "task-1350373" [ 1334.083939] env[63355]: _type = "Task" [ 1334.083939] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.091406] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350373, 'name': CreateVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.344639] env[63355]: DEBUG nova.compute.manager [req-6894c835-06d2-4e8e-915d-6d132c318d88 req-79dc7e56-527b-4dc6-9c13-b4beed742c38 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Received event network-changed-1f3d442f-7173-41c7-b933-3cddb214c854 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1334.344920] env[63355]: DEBUG nova.compute.manager [req-6894c835-06d2-4e8e-915d-6d132c318d88 req-79dc7e56-527b-4dc6-9c13-b4beed742c38 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Refreshing instance network info cache due to event network-changed-1f3d442f-7173-41c7-b933-3cddb214c854. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1334.345218] env[63355]: DEBUG oslo_concurrency.lockutils [req-6894c835-06d2-4e8e-915d-6d132c318d88 req-79dc7e56-527b-4dc6-9c13-b4beed742c38 service nova] Acquiring lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1334.345449] env[63355]: DEBUG oslo_concurrency.lockutils [req-6894c835-06d2-4e8e-915d-6d132c318d88 req-79dc7e56-527b-4dc6-9c13-b4beed742c38 service nova] Acquired lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1334.345703] env[63355]: DEBUG nova.network.neutron [req-6894c835-06d2-4e8e-915d-6d132c318d88 req-79dc7e56-527b-4dc6-9c13-b4beed742c38 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Refreshing network info cache for port 1f3d442f-7173-41c7-b933-3cddb214c854 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1334.594215] env[63355]: DEBUG oslo_vmware.api [-] Task: {'id': task-1350373, 'name': CreateVM_Task, 'duration_secs': 0.301082} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.594564] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Created VM on the ESX host {{(pid=63355) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1334.595072] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'mount_device': '/dev/sda', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287787', 'volume_id': 'f9514f73-1171-47d3-92bf-d2c1660b77d0', 'name': 'volume-f9514f73-1171-47d3-92bf-d2c1660b77d0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c2629183-3353-4bc6-8bf6-8fc72a874620', 'attached_at': '', 'detached_at': '', 'volume_id': 'f9514f73-1171-47d3-92bf-d2c1660b77d0', 'serial': 'f9514f73-1171-47d3-92bf-d2c1660b77d0'}, 'attachment_id': '6e927790-be9e-4e93-9ebc-d64583996bd7', 'guest_format': None, 'delete_on_termination': True, 'disk_bus': None, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=63355) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1334.595298] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Root volume attach. Driver type: vmdk {{(pid=63355) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1334.596108] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5cb1e6-e7d8-431c-ba15-89cb49ad61c7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.603094] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d4e456-9348-4330-b715-d22030187287 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.609226] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ed1864-f618-4a65-bf01-f0571dae8d95 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.615712] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-8c4a2d4f-b7e0-4a89-8abe-58bf3b69332d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.622412] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1334.622412] env[63355]: value = "task-1350374" [ 1334.622412] env[63355]: _type = "Task" [ 1334.622412] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.629341] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350374, 'name': RelocateVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.027834] env[63355]: DEBUG nova.network.neutron [req-6894c835-06d2-4e8e-915d-6d132c318d88 req-79dc7e56-527b-4dc6-9c13-b4beed742c38 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updated VIF entry in instance network info cache for port 1f3d442f-7173-41c7-b933-3cddb214c854. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1335.028270] env[63355]: DEBUG nova.network.neutron [req-6894c835-06d2-4e8e-915d-6d132c318d88 req-79dc7e56-527b-4dc6-9c13-b4beed742c38 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance_info_cache with network_info: [{"id": "1f3d442f-7173-41c7-b933-3cddb214c854", "address": "fa:16:3e:e8:8f:00", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f3d442f-71", "ovs_interfaceid": "1f3d442f-7173-41c7-b933-3cddb214c854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1335.132863] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350374, 'name': RelocateVM_Task, 'duration_secs': 0.319724} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.132863] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1335.132967] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287787', 'volume_id': 'f9514f73-1171-47d3-92bf-d2c1660b77d0', 'name': 'volume-f9514f73-1171-47d3-92bf-d2c1660b77d0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c2629183-3353-4bc6-8bf6-8fc72a874620', 'attached_at': '', 'detached_at': '', 'volume_id': 'f9514f73-1171-47d3-92bf-d2c1660b77d0', 'serial': 'f9514f73-1171-47d3-92bf-d2c1660b77d0'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1335.133731] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f0e2a6-ddd4-438e-9ef0-a1ec3d55457d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.148576] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a54763-7552-4a62-a052-822b336597b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.169400] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-f9514f73-1171-47d3-92bf-d2c1660b77d0/volume-f9514f73-1171-47d3-92bf-d2c1660b77d0.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1335.170066] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74282801-7516-4fa2-84f9-a6375507135e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.189964] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1335.189964] env[63355]: value = "task-1350375" [ 1335.189964] env[63355]: _type = "Task" [ 1335.189964] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.197426] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350375, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.530763] env[63355]: DEBUG oslo_concurrency.lockutils [req-6894c835-06d2-4e8e-915d-6d132c318d88 req-79dc7e56-527b-4dc6-9c13-b4beed742c38 service nova] Releasing lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1335.700132] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350375, 'name': ReconfigVM_Task, 'duration_secs': 0.238592} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.700525] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-f9514f73-1171-47d3-92bf-d2c1660b77d0/volume-f9514f73-1171-47d3-92bf-d2c1660b77d0.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1335.704948] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fedc25ad-27ae-4998-8a0d-0a24ba663d81 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.719605] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1335.719605] env[63355]: value = "task-1350376" [ 1335.719605] env[63355]: _type = "Task" [ 1335.719605] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.727722] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350376, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.229478] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350376, 'name': ReconfigVM_Task, 'duration_secs': 0.115689} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.229784] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287787', 'volume_id': 'f9514f73-1171-47d3-92bf-d2c1660b77d0', 'name': 'volume-f9514f73-1171-47d3-92bf-d2c1660b77d0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c2629183-3353-4bc6-8bf6-8fc72a874620', 'attached_at': '', 'detached_at': '', 'volume_id': 'f9514f73-1171-47d3-92bf-d2c1660b77d0', 'serial': 'f9514f73-1171-47d3-92bf-d2c1660b77d0'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1336.230336] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f63b0e4f-997b-4f87-aca7-af982b7dab06 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.236239] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1336.236239] env[63355]: value = "task-1350377" [ 1336.236239] env[63355]: _type = "Task" [ 1336.236239] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.246606] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350377, 'name': Rename_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.746146] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350377, 'name': Rename_Task, 'duration_secs': 0.137306} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.746464] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1336.746717] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d59b00da-8018-459e-bcaf-9bdb514ab6e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.752754] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1336.752754] env[63355]: value = "task-1350378" [ 1336.752754] env[63355]: _type = "Task" [ 1336.752754] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.760672] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350378, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.262933] env[63355]: DEBUG oslo_vmware.api [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350378, 'name': PowerOnVM_Task, 'duration_secs': 0.464403} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.263198] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1337.263404] env[63355]: INFO nova.compute.manager [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Took 3.38 seconds to spawn the instance on the hypervisor. [ 1337.263584] env[63355]: DEBUG nova.compute.manager [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Checking state {{(pid=63355) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1337.264331] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96883e6-7444-4995-a45e-523e79189908 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.784056] env[63355]: INFO nova.compute.manager [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Took 9.78 seconds to build instance. [ 1338.285753] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ed030f08-86cd-4226-b7f0-afb9c3bd1dd8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.288s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1339.009148] env[63355]: DEBUG nova.compute.manager [req-c5950197-ecc6-431f-954b-fae700e4c7cd req-99705202-8380-4950-9b6d-b6fc9d9a8468 service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Received event network-changed-e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1339.009148] env[63355]: DEBUG nova.compute.manager [req-c5950197-ecc6-431f-954b-fae700e4c7cd req-99705202-8380-4950-9b6d-b6fc9d9a8468 service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Refreshing instance network info cache due to event network-changed-e3455ba6-ed91-4901-b232-76ca673421a7. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1339.009148] env[63355]: DEBUG oslo_concurrency.lockutils [req-c5950197-ecc6-431f-954b-fae700e4c7cd req-99705202-8380-4950-9b6d-b6fc9d9a8468 service nova] Acquiring lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1339.009148] env[63355]: DEBUG oslo_concurrency.lockutils [req-c5950197-ecc6-431f-954b-fae700e4c7cd req-99705202-8380-4950-9b6d-b6fc9d9a8468 service nova] Acquired lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1339.010994] env[63355]: DEBUG nova.network.neutron [req-c5950197-ecc6-431f-954b-fae700e4c7cd req-99705202-8380-4950-9b6d-b6fc9d9a8468 service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Refreshing network info cache for port e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1339.765987] env[63355]: DEBUG nova.network.neutron [req-c5950197-ecc6-431f-954b-fae700e4c7cd req-99705202-8380-4950-9b6d-b6fc9d9a8468 service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updated VIF entry in instance network info cache for port e3455ba6-ed91-4901-b232-76ca673421a7. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1339.766383] env[63355]: DEBUG nova.network.neutron [req-c5950197-ecc6-431f-954b-fae700e4c7cd req-99705202-8380-4950-9b6d-b6fc9d9a8468 service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updating instance_info_cache with network_info: [{"id": "e3455ba6-ed91-4901-b232-76ca673421a7", "address": "fa:16:3e:eb:3b:b7", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3455ba6-ed", "ovs_interfaceid": "e3455ba6-ed91-4901-b232-76ca673421a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1340.269251] env[63355]: DEBUG oslo_concurrency.lockutils [req-c5950197-ecc6-431f-954b-fae700e4c7cd req-99705202-8380-4950-9b6d-b6fc9d9a8468 service nova] Releasing lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1340.524520] env[63355]: DEBUG nova.compute.manager [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Stashing vm_state: active {{(pid=63355) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1341.032873] env[63355]: DEBUG nova.compute.manager [req-17b61ceb-015c-4d3a-a1f9-825e2ad0c426 req-e24e8dbc-9d20-4109-835f-2128c4c665dd service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Received event network-changed-1f3d442f-7173-41c7-b933-3cddb214c854 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1341.033088] env[63355]: DEBUG nova.compute.manager [req-17b61ceb-015c-4d3a-a1f9-825e2ad0c426 req-e24e8dbc-9d20-4109-835f-2128c4c665dd service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Refreshing instance network info cache due to event network-changed-1f3d442f-7173-41c7-b933-3cddb214c854. {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1341.033305] env[63355]: DEBUG oslo_concurrency.lockutils [req-17b61ceb-015c-4d3a-a1f9-825e2ad0c426 req-e24e8dbc-9d20-4109-835f-2128c4c665dd service nova] Acquiring lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1341.033451] env[63355]: DEBUG oslo_concurrency.lockutils [req-17b61ceb-015c-4d3a-a1f9-825e2ad0c426 req-e24e8dbc-9d20-4109-835f-2128c4c665dd service nova] Acquired lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1341.033622] env[63355]: DEBUG nova.network.neutron [req-17b61ceb-015c-4d3a-a1f9-825e2ad0c426 req-e24e8dbc-9d20-4109-835f-2128c4c665dd service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Refreshing network info cache for port 1f3d442f-7173-41c7-b933-3cddb214c854 {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1341.044082] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1341.044330] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1341.548705] env[63355]: INFO nova.compute.claims [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1341.728664] env[63355]: DEBUG nova.network.neutron [req-17b61ceb-015c-4d3a-a1f9-825e2ad0c426 req-e24e8dbc-9d20-4109-835f-2128c4c665dd service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updated VIF entry in instance network info cache for port 1f3d442f-7173-41c7-b933-3cddb214c854. {{(pid=63355) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1341.729029] env[63355]: DEBUG nova.network.neutron [req-17b61ceb-015c-4d3a-a1f9-825e2ad0c426 req-e24e8dbc-9d20-4109-835f-2128c4c665dd service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance_info_cache with network_info: [{"id": "1f3d442f-7173-41c7-b933-3cddb214c854", "address": "fa:16:3e:e8:8f:00", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f3d442f-71", "ovs_interfaceid": "1f3d442f-7173-41c7-b933-3cddb214c854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1342.055358] env[63355]: INFO nova.compute.resource_tracker [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating resource usage from migration 0407e9cc-30e9-4f25-9ea4-a6d6832d1eb3 [ 1342.120483] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ec8093-7118-40a1-a931-bb40086b7f9b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.127697] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3a0a0b-34bd-4618-a4ec-a31971eca931 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.156058] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef69de4-28c0-4e07-9a48-0cedd82fa9cb {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.162505] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312d7b2b-25c4-4648-b3c8-705dc719d3a4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.174866] env[63355]: DEBUG nova.compute.provider_tree [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1342.231211] env[63355]: DEBUG oslo_concurrency.lockutils [req-17b61ceb-015c-4d3a-a1f9-825e2ad0c426 req-e24e8dbc-9d20-4109-835f-2128c4c665dd service nova] Releasing lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1342.678331] env[63355]: DEBUG nova.scheduler.client.report [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1343.183252] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.139s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1343.183471] env[63355]: INFO nova.compute.manager [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Migrating [ 1343.697363] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1343.697727] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1343.697802] env[63355]: DEBUG nova.network.neutron [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1344.389496] env[63355]: DEBUG nova.network.neutron [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance_info_cache with network_info: [{"id": "1f3d442f-7173-41c7-b933-3cddb214c854", "address": "fa:16:3e:e8:8f:00", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f3d442f-71", "ovs_interfaceid": "1f3d442f-7173-41c7-b933-3cddb214c854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1344.892677] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1346.406672] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d47a30-e298-4392-bb75-3898887982ac {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.425033] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance 'c2629183-3353-4bc6-8bf6-8fc72a874620' progress to 0 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1346.931574] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1346.931890] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f3122f9-84a2-4e15-8f6e-b5dd94a8e267 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.940172] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1346.940172] env[63355]: value = "task-1350379" [ 1346.940172] env[63355]: _type = "Task" [ 1346.940172] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.947974] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350379, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.450593] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350379, 'name': PowerOffVM_Task, 'duration_secs': 0.304815} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.451020] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1347.451081] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance 'c2629183-3353-4bc6-8bf6-8fc72a874620' progress to 17 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1347.958059] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1347.958351] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1347.958589] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1347.958838] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1347.959010] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1347.959170] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1347.959393] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1347.959610] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1347.959792] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1347.959963] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1347.960158] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1347.965308] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06d2d509-ec91-4b34-a03f-3003bd3129cf {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.985147] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1347.985147] env[63355]: value = "task-1350380" [ 1347.985147] env[63355]: _type = "Task" [ 1347.985147] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.992915] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350380, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.494845] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350380, 'name': ReconfigVM_Task, 'duration_secs': 0.189078} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.495256] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance 'c2629183-3353-4bc6-8bf6-8fc72a874620' progress to 33 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1349.001919] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T02:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1349.002222] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1349.002361] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image limits 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1349.002554] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Flavor pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1349.002706] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Image pref 0:0:0 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1349.002858] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63355) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1349.003082] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1349.003249] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1349.003426] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Got 1 possible topologies {{(pid=63355) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1349.003591] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1349.003765] env[63355]: DEBUG nova.virt.hardware [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63355) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1349.008947] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1349.009243] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fe5fa6b-7ad5-4f4b-b44b-5c0b2d3e686a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.028099] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1349.028099] env[63355]: value = "task-1350381" [ 1349.028099] env[63355]: _type = "Task" [ 1349.028099] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.035375] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350381, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.537653] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350381, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1350.038518] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350381, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1350.539334] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350381, 'name': ReconfigVM_Task, 'duration_secs': 1.132237} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1350.539684] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1350.540314] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5cdea84-432a-4cab-9228-16d504f4fd0d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.561065] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-f9514f73-1171-47d3-92bf-d2c1660b77d0/volume-f9514f73-1171-47d3-92bf-d2c1660b77d0.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1350.561295] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0596456f-569f-4546-ac2d-3cb0f3514255 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.577942] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1350.577942] env[63355]: value = "task-1350382" [ 1350.577942] env[63355]: _type = "Task" [ 1350.577942] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1350.585011] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350382, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1351.089183] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350382, 'name': ReconfigVM_Task, 'duration_secs': 0.266503} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1351.089470] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-f9514f73-1171-47d3-92bf-d2c1660b77d0/volume-f9514f73-1171-47d3-92bf-d2c1660b77d0.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1351.089734] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance 'c2629183-3353-4bc6-8bf6-8fc72a874620' progress to 50 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1351.596366] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01222f92-0032-43c4-a1e9-bccb95c5021c {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.259081] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd963c83-5570-4d4f-970a-85fc9b00915b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.276164] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance 'c2629183-3353-4bc6-8bf6-8fc72a874620' progress to 67 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1352.515022] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "9f7c978f-e993-4fa9-9402-489545a10020" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1352.515022] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1353.018308] env[63355]: DEBUG nova.compute.utils [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Using /dev/sd instead of None {{(pid=63355) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1353.520806] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1353.904408] env[63355]: DEBUG nova.network.neutron [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Port 1f3d442f-7173-41c7-b933-3cddb214c854 binding to destination host cpu-1 is already ACTIVE {{(pid=63355) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1354.577173] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "9f7c978f-e993-4fa9-9402-489545a10020" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1354.577562] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.577690] env[63355]: INFO nova.compute.manager [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Attaching volume 8eb71f15-0320-41ae-ad9d-8d4130b6d1a9 to /dev/sdb [ 1354.609045] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200b6fb0-2c55-44a6-ac16-2b7985f92365 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.616200] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23e29b1-86c2-4f59-bbb3-d8414a1ae0b5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.630584] env[63355]: DEBUG nova.virt.block_device [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Updating existing volume attachment record: b4722410-da4e-4ba3-848d-713f665bdcdd {{(pid=63355) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1354.926294] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "c2629183-3353-4bc6-8bf6-8fc72a874620-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1354.927547] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.927547] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1355.960935] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1355.961168] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1355.961342] env[63355]: DEBUG nova.network.neutron [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1356.658226] env[63355]: DEBUG nova.network.neutron [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance_info_cache with network_info: [{"id": "1f3d442f-7173-41c7-b933-3cddb214c854", "address": "fa:16:3e:e8:8f:00", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f3d442f-71", "ovs_interfaceid": "1f3d442f-7173-41c7-b933-3cddb214c854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1357.161048] env[63355]: DEBUG oslo_concurrency.lockutils [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1357.670095] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281163af-9c1c-4a8d-9b4f-4705f2726d67 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.677129] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550edb31-6ed6-4796-8cb8-6dfa8801e8af {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.783097] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc1e420-18cb-4123-8d4f-08d6423224ee {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.802942] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6195b9c5-bd5c-4cf7-ad08-f8e83d316d0e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.809241] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance 'c2629183-3353-4bc6-8bf6-8fc72a874620' progress to 83 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1359.172566] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Volume attach. Driver type: vmdk {{(pid=63355) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1359.172834] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287789', 'volume_id': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'name': 'volume-8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9f7c978f-e993-4fa9-9402-489545a10020', 'attached_at': '', 'detached_at': '', 'volume_id': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'serial': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1359.173727] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14792100-e8ae-4499-9461-2ee3e6c43d80 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.189753] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4328693a-78fc-4c35-a4b8-a2fcd301a815 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.213006] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] volume-8eb71f15-0320-41ae-ad9d-8d4130b6d1a9/volume-8eb71f15-0320-41ae-ad9d-8d4130b6d1a9.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1359.213246] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce2edf6d-b3f9-45b9-8c5f-ca0457486aaa {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.230294] env[63355]: DEBUG oslo_vmware.api [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1359.230294] env[63355]: value = "task-1350385" [ 1359.230294] env[63355]: _type = "Task" [ 1359.230294] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.237323] env[63355]: DEBUG oslo_vmware.api [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350385, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.315580] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Powering on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1359.315877] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7fde64f-4f4b-40c1-bdff-dba95653d265 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.322086] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1359.322086] env[63355]: value = "task-1350386" [ 1359.322086] env[63355]: _type = "Task" [ 1359.322086] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.329200] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350386, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.740149] env[63355]: DEBUG oslo_vmware.api [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350385, 'name': ReconfigVM_Task, 'duration_secs': 0.325872} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.740448] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Reconfigured VM instance instance-0000006f to attach disk [datastore2] volume-8eb71f15-0320-41ae-ad9d-8d4130b6d1a9/volume-8eb71f15-0320-41ae-ad9d-8d4130b6d1a9.vmdk or device None with type thin {{(pid=63355) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1359.744928] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a59953c6-22e3-42c0-ab20-5462679e32f3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.759319] env[63355]: DEBUG oslo_vmware.api [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1359.759319] env[63355]: value = "task-1350387" [ 1359.759319] env[63355]: _type = "Task" [ 1359.759319] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.767055] env[63355]: DEBUG oslo_vmware.api [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350387, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.831613] env[63355]: DEBUG oslo_vmware.api [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350386, 'name': PowerOnVM_Task, 'duration_secs': 0.345771} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.831959] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Powered on the VM {{(pid=63355) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1359.832074] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-2fd2a715-f661-448f-91c7-2fc05e5c42f8 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance 'c2629183-3353-4bc6-8bf6-8fc72a874620' progress to 100 {{(pid=63355) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1360.268968] env[63355]: DEBUG oslo_vmware.api [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350387, 'name': ReconfigVM_Task, 'duration_secs': 0.14465} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.269338] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287789', 'volume_id': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'name': 'volume-8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9f7c978f-e993-4fa9-9402-489545a10020', 'attached_at': '', 'detached_at': '', 'volume_id': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'serial': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9'} {{(pid=63355) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1361.305444] env[63355]: DEBUG nova.objects.instance [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lazy-loading 'flavor' on Instance uuid 9f7c978f-e993-4fa9-9402-489545a10020 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1361.811139] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e0ca3b28-0f87-482b-9e42-c42843874ffa tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.233s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1361.903188] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "c2629183-3353-4bc6-8bf6-8fc72a874620" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1361.903514] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1361.903638] env[63355]: DEBUG nova.compute.manager [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Going to confirm migration 6 {{(pid=63355) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1362.062238] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "9f7c978f-e993-4fa9-9402-489545a10020" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1362.062536] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1362.472296] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1362.472532] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquired lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1362.472667] env[63355]: DEBUG nova.network.neutron [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Building network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1362.472852] env[63355]: DEBUG nova.objects.instance [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lazy-loading 'info_cache' on Instance uuid c2629183-3353-4bc6-8bf6-8fc72a874620 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1362.566795] env[63355]: INFO nova.compute.manager [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Detaching volume 8eb71f15-0320-41ae-ad9d-8d4130b6d1a9 [ 1362.600035] env[63355]: INFO nova.virt.block_device [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Attempting to driver detach volume 8eb71f15-0320-41ae-ad9d-8d4130b6d1a9 from mountpoint /dev/sdb [ 1362.600298] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1362.600485] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287789', 'volume_id': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'name': 'volume-8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9f7c978f-e993-4fa9-9402-489545a10020', 'attached_at': '', 'detached_at': '', 'volume_id': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'serial': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1362.601415] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e524b84-1b00-4363-bd94-f615b1b5c5a7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.622380] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22dfe887-7de7-4ad6-9254-903e39caf6fd {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.630033] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557d83d3-903c-40d5-b1f1-be0ab6efae9e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.649144] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63ebf3d-9725-4177-9228-99b0be9b2b86 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.662909] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] The volume has not been displaced from its original location: [datastore2] volume-8eb71f15-0320-41ae-ad9d-8d4130b6d1a9/volume-8eb71f15-0320-41ae-ad9d-8d4130b6d1a9.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1362.667938] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1362.668228] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efa6e09f-c619-43e5-836d-81b8fffadf95 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.686178] env[63355]: DEBUG oslo_vmware.api [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1362.686178] env[63355]: value = "task-1350388" [ 1362.686178] env[63355]: _type = "Task" [ 1362.686178] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.693556] env[63355]: DEBUG oslo_vmware.api [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350388, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.196717] env[63355]: DEBUG oslo_vmware.api [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350388, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.678862] env[63355]: DEBUG nova.network.neutron [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance_info_cache with network_info: [{"id": "1f3d442f-7173-41c7-b933-3cddb214c854", "address": "fa:16:3e:e8:8f:00", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f3d442f-71", "ovs_interfaceid": "1f3d442f-7173-41c7-b933-3cddb214c854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1363.697245] env[63355]: DEBUG oslo_vmware.api [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350388, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.181873] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Releasing lock "refresh_cache-c2629183-3353-4bc6-8bf6-8fc72a874620" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1364.182176] env[63355]: DEBUG nova.objects.instance [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lazy-loading 'migration_context' on Instance uuid c2629183-3353-4bc6-8bf6-8fc72a874620 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1364.197384] env[63355]: DEBUG oslo_vmware.api [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350388, 'name': ReconfigVM_Task, 'duration_secs': 1.271725} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.197633] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1364.202345] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28113492-f1f0-43e9-be6e-8ee99f28557e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.216629] env[63355]: DEBUG oslo_vmware.api [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1364.216629] env[63355]: value = "task-1350389" [ 1364.216629] env[63355]: _type = "Task" [ 1364.216629] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.223677] env[63355]: DEBUG oslo_vmware.api [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350389, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.685849] env[63355]: DEBUG nova.objects.base [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63355) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1364.686859] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30b7dc2-32d0-42f3-95cb-80e642f7f2e7 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.706598] env[63355]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e364e3e-2cd0-43b9-aeea-48c60fb1f3af {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.711847] env[63355]: DEBUG oslo_vmware.api [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1364.711847] env[63355]: value = "session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5259e334-f5f9-8b55-d390-757d59c5a002" [ 1364.711847] env[63355]: _type = "Task" [ 1364.711847] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.719233] env[63355]: DEBUG oslo_vmware.api [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5259e334-f5f9-8b55-d390-757d59c5a002, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.726049] env[63355]: DEBUG oslo_vmware.api [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350389, 'name': ReconfigVM_Task, 'duration_secs': 0.131345} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.726328] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287789', 'volume_id': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'name': 'volume-8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9f7c978f-e993-4fa9-9402-489545a10020', 'attached_at': '', 'detached_at': '', 'volume_id': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9', 'serial': '8eb71f15-0320-41ae-ad9d-8d4130b6d1a9'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1365.222028] env[63355]: DEBUG oslo_vmware.api [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': session[52d3e5a7-a04a-f4ae-e518-aa486c88d1a0]5259e334-f5f9-8b55-d390-757d59c5a002, 'name': SearchDatastore_Task, 'duration_secs': 0.008236} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.222337] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1365.222569] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1365.266450] env[63355]: DEBUG nova.objects.instance [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lazy-loading 'flavor' on Instance uuid 9f7c978f-e993-4fa9-9402-489545a10020 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1365.794207] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381d1cdf-5e75-48d0-a1b3-3a0b6e5cbe5b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.801201] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843b319e-0ad0-4a0b-a5c7-eb9675d994c0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.829946] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31e171f-8e40-49d2-96c2-2b35ef824389 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.836638] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7951d7bf-632d-448e-b871-ade03ce424f0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.849206] env[63355]: DEBUG nova.compute.provider_tree [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1366.273256] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47668e30-e0cf-4263-ab92-45128292b1fc tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.211s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1366.352286] env[63355]: DEBUG nova.scheduler.client.report [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1367.333225] env[63355]: DEBUG oslo_concurrency.lockutils [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "9f7c978f-e993-4fa9-9402-489545a10020" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1367.333627] env[63355]: DEBUG oslo_concurrency.lockutils [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1367.333678] env[63355]: DEBUG oslo_concurrency.lockutils [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "9f7c978f-e993-4fa9-9402-489545a10020-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1367.333832] env[63355]: DEBUG oslo_concurrency.lockutils [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1367.334019] env[63355]: DEBUG oslo_concurrency.lockutils [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1367.336175] env[63355]: INFO nova.compute.manager [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Terminating instance [ 1367.337940] env[63355]: DEBUG nova.compute.manager [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1367.338152] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1367.338995] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcb2a3f-a868-4679-ba54-6b659994512b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.346769] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1367.346983] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db040a9f-cdd0-453e-a20c-ab6255222651 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.352931] env[63355]: DEBUG oslo_vmware.api [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1367.352931] env[63355]: value = "task-1350390" [ 1367.352931] env[63355]: _type = "Task" [ 1367.352931] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.359814] env[63355]: DEBUG oslo_vmware.api [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350390, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.360913] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.138s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1367.862424] env[63355]: DEBUG oslo_vmware.api [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350390, 'name': PowerOffVM_Task, 'duration_secs': 0.223396} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.862690] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1367.862860] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1367.863167] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-779328c2-a501-4c6a-849e-4008bb4e8f84 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.912819] env[63355]: INFO nova.scheduler.client.report [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted allocation for migration 0407e9cc-30e9-4f25-9ea4-a6d6832d1eb3 [ 1367.931853] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1367.931853] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1367.931853] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleting the datastore file [datastore2] 9f7c978f-e993-4fa9-9402-489545a10020 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1367.931853] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c10eb826-8638-448e-bad3-669b17eab515 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.939292] env[63355]: DEBUG oslo_vmware.api [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for the task: (returnval){ [ 1367.939292] env[63355]: value = "task-1350392" [ 1367.939292] env[63355]: _type = "Task" [ 1367.939292] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.947510] env[63355]: DEBUG oslo_vmware.api [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350392, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.258710] env[63355]: INFO nova.compute.manager [None req-b203bf82-9170-4e6f-8717-58f8b62b891d tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Get console output [ 1368.259352] env[63355]: WARNING nova.virt.vmwareapi.driver [None req-b203bf82-9170-4e6f-8717-58f8b62b891d tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] The console log is missing. Check your VSPC configuration [ 1368.418611] env[63355]: DEBUG oslo_concurrency.lockutils [None req-47424961-f0a0-4f92-bf04-9969b8489430 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.515s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1368.449062] env[63355]: DEBUG oslo_vmware.api [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Task: {'id': task-1350392, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148866} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1368.449337] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1368.449480] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1368.449655] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1368.449872] env[63355]: INFO nova.compute.manager [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1368.450133] env[63355]: DEBUG oslo.service.loopingcall [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1368.450325] env[63355]: DEBUG nova.compute.manager [-] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1368.450420] env[63355]: DEBUG nova.network.neutron [-] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1368.870185] env[63355]: DEBUG nova.compute.manager [req-caebdee9-1d24-4291-ac8c-fa04bae87302 req-cc7b5802-2cfd-4e05-8fd3-ddddd77c6827 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Received event network-vif-deleted-70f5b776-10c7-4fa8-be9e-a318da6adf88 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1368.870185] env[63355]: INFO nova.compute.manager [req-caebdee9-1d24-4291-ac8c-fa04bae87302 req-cc7b5802-2cfd-4e05-8fd3-ddddd77c6827 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Neutron deleted interface 70f5b776-10c7-4fa8-be9e-a318da6adf88; detaching it from the instance and deleting it from the info cache [ 1368.870185] env[63355]: DEBUG nova.network.neutron [req-caebdee9-1d24-4291-ac8c-fa04bae87302 req-cc7b5802-2cfd-4e05-8fd3-ddddd77c6827 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1369.349060] env[63355]: DEBUG nova.network.neutron [-] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1369.373630] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1222901-c7e7-4e6e-9f3e-ee5771405e25 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.383975] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2025d5-5981-4f22-8179-133cc89b04c4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.409206] env[63355]: DEBUG nova.compute.manager [req-caebdee9-1d24-4291-ac8c-fa04bae87302 req-cc7b5802-2cfd-4e05-8fd3-ddddd77c6827 service nova] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Detach interface failed, port_id=70f5b776-10c7-4fa8-be9e-a318da6adf88, reason: Instance 9f7c978f-e993-4fa9-9402-489545a10020 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1369.852022] env[63355]: INFO nova.compute.manager [-] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Took 1.40 seconds to deallocate network for instance. [ 1370.133691] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1370.359031] env[63355]: DEBUG oslo_concurrency.lockutils [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1370.359405] env[63355]: DEBUG oslo_concurrency.lockutils [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1370.359673] env[63355]: DEBUG nova.objects.instance [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lazy-loading 'resources' on Instance uuid 9f7c978f-e993-4fa9-9402-489545a10020 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1370.637085] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1370.926711] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883297b8-33a7-4e34-895f-b1fc4e9f8a6e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.935233] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3183d1-0515-4e6d-9ec5-b3b317b0a1c1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.970815] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c16f41-498a-461c-8885-0a8012ecc6d9 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.978746] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e1f540-2c6e-460d-8be7-ced5afd4bac0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.992316] env[63355]: DEBUG nova.compute.provider_tree [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1371.495738] env[63355]: DEBUG nova.scheduler.client.report [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1372.000847] env[63355]: DEBUG oslo_concurrency.lockutils [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.003136] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.366s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.003319] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.003476] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1372.004332] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9c1237-997c-44b4-bd63-d5b7a57a494e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.012397] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6814c27b-bf53-4989-9c6e-0203f74ba392 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.025981] env[63355]: INFO nova.scheduler.client.report [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Deleted allocations for instance 9f7c978f-e993-4fa9-9402-489545a10020 [ 1372.027455] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c7bdee-ce01-4e74-ac26-ceec3f43c7e3 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.035590] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89adc218-d36a-46e1-a853-29b7391db883 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.064415] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181079MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1372.064584] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1372.064764] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.535621] env[63355]: DEBUG oslo_concurrency.lockutils [None req-586c0f4c-51df-4b7e-ac6f-00bb9dfa99b2 tempest-AttachVolumeNegativeTest-660443579 tempest-AttachVolumeNegativeTest-660443579-project-member] Lock "9f7c978f-e993-4fa9-9402-489545a10020" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.202s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1373.085163] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 8db23054-4556-496a-a4b1-a24c71c30f50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1373.085414] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance 09e632ab-cd73-48cd-a4f3-308c247794c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1373.085467] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Instance c2629183-3353-4bc6-8bf6-8fc72a874620 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63355) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1373.085611] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1373.085744] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1152MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1373.126233] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be04ff8c-5422-40de-970e-a0f5c82c5a46 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.133774] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9750170-cd53-4097-be46-813b627c0931 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.162388] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1b813f-318b-4d6b-b116-bbb86a569eb0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.169345] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136110d9-de3b-406a-803c-068e0da21352 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.181764] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1373.685381] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1374.190480] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1374.190905] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.126s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1379.186572] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1379.186882] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1379.186937] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1379.187068] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Rebuilding the list of instances to heal {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1379.716681] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1379.716842] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquired lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1379.716993] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Forcefully refreshing network info cache for instance {{(pid=63355) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1379.717157] env[63355]: DEBUG nova.objects.instance [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lazy-loading 'info_cache' on Instance uuid 8db23054-4556-496a-a4b1-a24c71c30f50 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1381.597480] env[63355]: DEBUG nova.network.neutron [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updating instance_info_cache with network_info: [{"id": "e3455ba6-ed91-4901-b232-76ca673421a7", "address": "fa:16:3e:eb:3b:b7", "network": {"id": "519dd36d-a162-45bc-8cb5-7fe0ca7487b0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-523187208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff5b796e7bb0458dbd3b862dcae7bdd1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3455ba6-ed", "ovs_interfaceid": "e3455ba6-ed91-4901-b232-76ca673421a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1382.100504] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Releasing lock "refresh_cache-8db23054-4556-496a-a4b1-a24c71c30f50" {{(pid=63355) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1382.100712] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updated the network info_cache for instance {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1382.100970] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.101262] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.101487] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.101652] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.101799] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.101950] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.102126] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1387.044558] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1395.175149] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "c2629183-3353-4bc6-8bf6-8fc72a874620" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1395.175537] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1395.175679] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "c2629183-3353-4bc6-8bf6-8fc72a874620-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1395.175869] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1395.176055] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1395.179546] env[63355]: INFO nova.compute.manager [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Terminating instance [ 1395.181595] env[63355]: DEBUG nova.compute.manager [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1395.181828] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1395.182137] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-561bb65d-3783-4376-89fa-edeb7c0ebb7b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.189426] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1395.189426] env[63355]: value = "task-1350396" [ 1395.189426] env[63355]: _type = "Task" [ 1395.189426] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1395.198659] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350396, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1395.699198] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350396, 'name': PowerOffVM_Task, 'duration_secs': 0.163819} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1395.699461] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1395.699662] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Volume detach. Driver type: vmdk {{(pid=63355) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1395.699862] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287787', 'volume_id': 'f9514f73-1171-47d3-92bf-d2c1660b77d0', 'name': 'volume-f9514f73-1171-47d3-92bf-d2c1660b77d0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'c2629183-3353-4bc6-8bf6-8fc72a874620', 'attached_at': '2024-10-24T02:37:42.000000', 'detached_at': '', 'volume_id': 'f9514f73-1171-47d3-92bf-d2c1660b77d0', 'serial': 'f9514f73-1171-47d3-92bf-d2c1660b77d0'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1395.700604] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a03d3a-1739-4e4d-bc1e-b8459cadf400 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.719246] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed5c217-e0bf-430e-a9cc-9b57e7dcd083 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.725039] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ea7349-825b-419d-a528-5fad36fbc00f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.741528] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ec3d4c-fc9e-4f3f-91ff-dfeadf9a4c2a {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.755112] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] The volume has not been displaced from its original location: [datastore1] volume-f9514f73-1171-47d3-92bf-d2c1660b77d0/volume-f9514f73-1171-47d3-92bf-d2c1660b77d0.vmdk. No consolidation needed. {{(pid=63355) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1395.760113] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1395.760366] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76f88276-b783-47c9-9405-59b2873c19b8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.777151] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1395.777151] env[63355]: value = "task-1350397" [ 1395.777151] env[63355]: _type = "Task" [ 1395.777151] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1395.784767] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350397, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1396.287016] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350397, 'name': ReconfigVM_Task, 'duration_secs': 0.157735} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1396.287575] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=63355) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1396.291883] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bad7635b-fa65-4b4d-a5c5-87fa88a1f4ea {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.306209] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1396.306209] env[63355]: value = "task-1350398" [ 1396.306209] env[63355]: _type = "Task" [ 1396.306209] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.316182] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350398, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1396.816323] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350398, 'name': ReconfigVM_Task, 'duration_secs': 0.126735} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1396.816618] env[63355]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-287787', 'volume_id': 'f9514f73-1171-47d3-92bf-d2c1660b77d0', 'name': 'volume-f9514f73-1171-47d3-92bf-d2c1660b77d0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'c2629183-3353-4bc6-8bf6-8fc72a874620', 'attached_at': '2024-10-24T02:37:42.000000', 'detached_at': '', 'volume_id': 'f9514f73-1171-47d3-92bf-d2c1660b77d0', 'serial': 'f9514f73-1171-47d3-92bf-d2c1660b77d0'} {{(pid=63355) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1396.816902] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1396.817684] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e77f2eb-ce90-49a0-b197-b9918e387ae1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.823853] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1396.824076] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-526d25d3-3334-424d-80e7-a0bf89c606b0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.885293] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1396.885521] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1396.885704] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleting the datastore file [datastore1] c2629183-3353-4bc6-8bf6-8fc72a874620 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1396.885965] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-692ec803-5f53-4b6c-9848-890474d7dfab {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.892244] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1396.892244] env[63355]: value = "task-1350400" [ 1396.892244] env[63355]: _type = "Task" [ 1396.892244] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.899658] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.402375] env[63355]: DEBUG oslo_vmware.api [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.078871} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.402759] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1397.402798] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1397.402976] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1397.403176] env[63355]: INFO nova.compute.manager [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1397.403412] env[63355]: DEBUG oslo.service.loopingcall [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1397.403614] env[63355]: DEBUG nova.compute.manager [-] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1397.403709] env[63355]: DEBUG nova.network.neutron [-] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1397.832763] env[63355]: DEBUG nova.compute.manager [req-d3ab61e5-7b20-41c8-8a7a-dc44fe816243 req-4da5a345-dbbe-411c-bba0-53b79eb8c399 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Received event network-vif-deleted-1f3d442f-7173-41c7-b933-3cddb214c854 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1397.832954] env[63355]: INFO nova.compute.manager [req-d3ab61e5-7b20-41c8-8a7a-dc44fe816243 req-4da5a345-dbbe-411c-bba0-53b79eb8c399 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Neutron deleted interface 1f3d442f-7173-41c7-b933-3cddb214c854; detaching it from the instance and deleting it from the info cache [ 1397.833102] env[63355]: DEBUG nova.network.neutron [req-d3ab61e5-7b20-41c8-8a7a-dc44fe816243 req-4da5a345-dbbe-411c-bba0-53b79eb8c399 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1398.313436] env[63355]: DEBUG nova.network.neutron [-] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1398.335390] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c52a751c-11a6-47aa-9b17-4be494ddbc34 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.345112] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd1fb42-a8ad-413a-abdf-fe68bd57776f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.368183] env[63355]: DEBUG nova.compute.manager [req-d3ab61e5-7b20-41c8-8a7a-dc44fe816243 req-4da5a345-dbbe-411c-bba0-53b79eb8c399 service nova] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Detach interface failed, port_id=1f3d442f-7173-41c7-b933-3cddb214c854, reason: Instance c2629183-3353-4bc6-8bf6-8fc72a874620 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1398.816567] env[63355]: INFO nova.compute.manager [-] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Took 1.41 seconds to deallocate network for instance. [ 1399.360287] env[63355]: INFO nova.compute.manager [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Took 0.54 seconds to detach 1 volumes for instance. [ 1399.362489] env[63355]: DEBUG nova.compute.manager [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Deleting volume: f9514f73-1171-47d3-92bf-d2c1660b77d0 {{(pid=63355) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1399.898547] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1399.898958] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1399.899131] env[63355]: DEBUG nova.objects.instance [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lazy-loading 'resources' on Instance uuid c2629183-3353-4bc6-8bf6-8fc72a874620 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1400.452225] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cb10fa-86fc-4e12-8801-869ab4872653 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.459758] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ab0ec2-82b3-4c9a-93e4-2621fcb5c392 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.489615] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963d9cf9-b65d-4bcd-917c-f27fdc4bec02 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.496518] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58c14cf-ada1-4afe-b17f-1461c7a2a4ca {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.509012] env[63355]: DEBUG nova.compute.provider_tree [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1401.012546] env[63355]: DEBUG nova.scheduler.client.report [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1401.517805] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1401.534604] env[63355]: INFO nova.scheduler.client.report [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted allocations for instance c2629183-3353-4bc6-8bf6-8fc72a874620 [ 1402.043186] env[63355]: DEBUG oslo_concurrency.lockutils [None req-8cf18d56-0a13-4057-81f7-467882cdab68 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "c2629183-3353-4bc6-8bf6-8fc72a874620" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.868s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1402.826530] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "09e632ab-cd73-48cd-a4f3-308c247794c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1402.826797] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "09e632ab-cd73-48cd-a4f3-308c247794c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1402.827023] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "09e632ab-cd73-48cd-a4f3-308c247794c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1402.827216] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "09e632ab-cd73-48cd-a4f3-308c247794c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1402.827394] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "09e632ab-cd73-48cd-a4f3-308c247794c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1402.829544] env[63355]: INFO nova.compute.manager [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Terminating instance [ 1402.831332] env[63355]: DEBUG nova.compute.manager [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1402.831529] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1402.832380] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59acd40-33f2-44f1-bdf7-c8ff11eab966 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.840009] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1402.840236] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9740d8ea-45ff-4a4b-81f3-1fd450c1b507 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.846405] env[63355]: DEBUG oslo_vmware.api [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1402.846405] env[63355]: value = "task-1350402" [ 1402.846405] env[63355]: _type = "Task" [ 1402.846405] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1402.854375] env[63355]: DEBUG oslo_vmware.api [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350402, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1403.356360] env[63355]: DEBUG oslo_vmware.api [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350402, 'name': PowerOffVM_Task, 'duration_secs': 0.195207} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1403.356776] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1403.356776] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1403.356910] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97ae1795-7cb2-48c1-8995-c216dcf31cb4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.427541] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1403.427721] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Deleting contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1403.427890] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleting the datastore file [datastore1] 09e632ab-cd73-48cd-a4f3-308c247794c8 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1403.428174] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78811cd0-502f-4ab7-acbc-42672d1cc97d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.434213] env[63355]: DEBUG oslo_vmware.api [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1403.434213] env[63355]: value = "task-1350404" [ 1403.434213] env[63355]: _type = "Task" [ 1403.434213] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1403.441171] env[63355]: DEBUG oslo_vmware.api [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350404, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1403.944906] env[63355]: DEBUG oslo_vmware.api [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350404, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131736} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1403.945353] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1403.945541] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Deleted contents of the VM from datastore datastore1 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1403.945729] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1403.945901] env[63355]: INFO nova.compute.manager [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1403.946479] env[63355]: DEBUG oslo.service.loopingcall [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1403.946479] env[63355]: DEBUG nova.compute.manager [-] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1403.946645] env[63355]: DEBUG nova.network.neutron [-] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1404.212199] env[63355]: DEBUG nova.compute.manager [req-feb52d1a-d483-450a-82ce-dafdfe849a05 req-b177f26c-60c3-4451-9614-e8b69fbd5a34 service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Received event network-vif-deleted-7cec8c38-1693-4dde-b2b2-1165d43964e4 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1404.212337] env[63355]: INFO nova.compute.manager [req-feb52d1a-d483-450a-82ce-dafdfe849a05 req-b177f26c-60c3-4451-9614-e8b69fbd5a34 service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Neutron deleted interface 7cec8c38-1693-4dde-b2b2-1165d43964e4; detaching it from the instance and deleting it from the info cache [ 1404.212516] env[63355]: DEBUG nova.network.neutron [req-feb52d1a-d483-450a-82ce-dafdfe849a05 req-b177f26c-60c3-4451-9614-e8b69fbd5a34 service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1404.686920] env[63355]: DEBUG nova.network.neutron [-] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1404.715151] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4042e18-1544-4187-a37c-ccf6bf5fdb3f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.724622] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f74cdc3-2724-4b9f-9f66-47a039abb461 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.747344] env[63355]: DEBUG nova.compute.manager [req-feb52d1a-d483-450a-82ce-dafdfe849a05 req-b177f26c-60c3-4451-9614-e8b69fbd5a34 service nova] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Detach interface failed, port_id=7cec8c38-1693-4dde-b2b2-1165d43964e4, reason: Instance 09e632ab-cd73-48cd-a4f3-308c247794c8 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1405.190177] env[63355]: INFO nova.compute.manager [-] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Took 1.24 seconds to deallocate network for instance. [ 1405.696873] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1405.697198] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1405.697393] env[63355]: DEBUG nova.objects.instance [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lazy-loading 'resources' on Instance uuid 09e632ab-cd73-48cd-a4f3-308c247794c8 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1406.242215] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66acbaf-2164-4fb4-9c18-adb303edd31e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.249888] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e6ad9b-0a45-4e89-9b93-9f8a26e1a4d1 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.279654] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3df7d4-1f12-41ee-ac4f-fcbe5813001d {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.286447] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9448b8-e26e-43fe-99df-8644fee68d30 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.298936] env[63355]: DEBUG nova.compute.provider_tree [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1406.802066] env[63355]: DEBUG nova.scheduler.client.report [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1407.307432] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1407.325188] env[63355]: INFO nova.scheduler.client.report [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted allocations for instance 09e632ab-cd73-48cd-a4f3-308c247794c8 [ 1407.833825] env[63355]: DEBUG oslo_concurrency.lockutils [None req-ace597a1-c872-4d07-9744-87f067e4729a tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "09e632ab-cd73-48cd-a4f3-308c247794c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.007s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1411.399465] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "8db23054-4556-496a-a4b1-a24c71c30f50" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1411.399898] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "8db23054-4556-496a-a4b1-a24c71c30f50" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1411.399941] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "8db23054-4556-496a-a4b1-a24c71c30f50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1411.400159] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "8db23054-4556-496a-a4b1-a24c71c30f50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1411.400330] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "8db23054-4556-496a-a4b1-a24c71c30f50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1411.402581] env[63355]: INFO nova.compute.manager [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Terminating instance [ 1411.404348] env[63355]: DEBUG nova.compute.manager [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Start destroying the instance on the hypervisor. {{(pid=63355) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1411.404547] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Destroying instance {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1411.405373] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b804a5-c55e-4449-beb3-bddc48848560 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.412932] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Powering off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1411.413385] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9bdeadf8-4ab4-4c2e-818b-a0f2f9b2e039 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.419429] env[63355]: DEBUG oslo_vmware.api [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1411.419429] env[63355]: value = "task-1350406" [ 1411.419429] env[63355]: _type = "Task" [ 1411.419429] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.427196] env[63355]: DEBUG oslo_vmware.api [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350406, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.930350] env[63355]: DEBUG oslo_vmware.api [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350406, 'name': PowerOffVM_Task, 'duration_secs': 0.169603} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1411.930660] env[63355]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Powered off the VM {{(pid=63355) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1411.930865] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Unregistering the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1411.931160] env[63355]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5f0a9fa-0a82-4db3-b09f-2f6652732023 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.992780] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Unregistered the VM {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1411.992988] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Deleting contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1411.993192] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleting the datastore file [datastore2] 8db23054-4556-496a-a4b1-a24c71c30f50 {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1411.993496] env[63355]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eab0aad7-6ce1-46c6-9957-c95b85f341e0 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.999856] env[63355]: DEBUG oslo_vmware.api [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for the task: (returnval){ [ 1411.999856] env[63355]: value = "task-1350408" [ 1411.999856] env[63355]: _type = "Task" [ 1411.999856] env[63355]: } to complete. {{(pid=63355) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1412.007372] env[63355]: DEBUG oslo_vmware.api [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350408, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1412.509317] env[63355]: DEBUG oslo_vmware.api [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Task: {'id': task-1350408, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127804} completed successfully. {{(pid=63355) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1412.509702] env[63355]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted the datastore file {{(pid=63355) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1412.509760] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Deleted contents of the VM from datastore datastore2 {{(pid=63355) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1412.509937] env[63355]: DEBUG nova.virt.vmwareapi.vmops [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Instance destroyed {{(pid=63355) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1412.510207] env[63355]: INFO nova.compute.manager [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1412.510464] env[63355]: DEBUG oslo.service.loopingcall [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63355) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1412.510658] env[63355]: DEBUG nova.compute.manager [-] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Deallocating network for instance {{(pid=63355) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1412.510751] env[63355]: DEBUG nova.network.neutron [-] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] deallocate_for_instance() {{(pid=63355) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1412.753328] env[63355]: DEBUG nova.compute.manager [req-67f2eb86-c999-4a9e-b920-c333be5ce4d9 req-42f2d9e2-0996-45b7-9e2b-e556dc94eb4c service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Received event network-vif-deleted-e3455ba6-ed91-4901-b232-76ca673421a7 {{(pid=63355) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1412.753560] env[63355]: INFO nova.compute.manager [req-67f2eb86-c999-4a9e-b920-c333be5ce4d9 req-42f2d9e2-0996-45b7-9e2b-e556dc94eb4c service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Neutron deleted interface e3455ba6-ed91-4901-b232-76ca673421a7; detaching it from the instance and deleting it from the info cache [ 1412.753733] env[63355]: DEBUG nova.network.neutron [req-67f2eb86-c999-4a9e-b920-c333be5ce4d9 req-42f2d9e2-0996-45b7-9e2b-e556dc94eb4c service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1413.232476] env[63355]: DEBUG nova.network.neutron [-] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Updating instance_info_cache with network_info: [] {{(pid=63355) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1413.256723] env[63355]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-766a4a76-86c1-4879-a136-0a58ee9b9dde {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.266698] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad381456-9cca-489f-b98c-fb6ae8869ee8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.287734] env[63355]: DEBUG nova.compute.manager [req-67f2eb86-c999-4a9e-b920-c333be5ce4d9 req-42f2d9e2-0996-45b7-9e2b-e556dc94eb4c service nova] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Detach interface failed, port_id=e3455ba6-ed91-4901-b232-76ca673421a7, reason: Instance 8db23054-4556-496a-a4b1-a24c71c30f50 could not be found. {{(pid=63355) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1413.735306] env[63355]: INFO nova.compute.manager [-] [instance: 8db23054-4556-496a-a4b1-a24c71c30f50] Took 1.22 seconds to deallocate network for instance. [ 1414.241608] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1414.241896] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1414.242135] env[63355]: DEBUG nova.objects.instance [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lazy-loading 'resources' on Instance uuid 8db23054-4556-496a-a4b1-a24c71c30f50 {{(pid=63355) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1414.775587] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a540e4-01be-4f1c-9849-04c06ff5c3b6 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.783177] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a07e421-b50d-4f33-8e95-9f4f9421b248 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.812467] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d41beae-5cb7-407a-a15c-4eee35088ad5 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.819544] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1079cdcd-4c9c-468a-b5cc-ef000319a82e {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.832095] env[63355]: DEBUG nova.compute.provider_tree [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1415.335865] env[63355]: DEBUG nova.scheduler.client.report [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1415.842175] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1415.859140] env[63355]: INFO nova.scheduler.client.report [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Deleted allocations for instance 8db23054-4556-496a-a4b1-a24c71c30f50 [ 1416.366661] env[63355]: DEBUG oslo_concurrency.lockutils [None req-e8b1995a-a3db-42e7-bc67-f5f6d9702010 tempest-ServerActionsTestOtherA-2121844102 tempest-ServerActionsTestOtherA-2121844102-project-member] Lock "8db23054-4556-496a-a4b1-a24c71c30f50" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.967s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1430.133078] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1430.636194] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1430.636485] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1430.636680] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1430.636845] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63355) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1430.637770] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64192cc9-fe10-4874-9bf3-3154f3cc378f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.646196] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50a3796-b7da-4588-8f78-5ccffde2dd38 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.659944] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f40b2f-8ffd-49cf-befc-1527581042e4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.666016] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a49f87-cfff-4728-a460-26f07b932fc8 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.694705] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180914MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=63355) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1430.694858] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1430.695055] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1431.714803] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1431.715050] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63355) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1431.727547] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33193a26-780f-4851-8a0e-20ce113b047f {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.734582] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed0a4aa-50ec-4ebb-966f-4a60286d3aef {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.762573] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31e36c9-ec19-4126-9df1-ad3107204bc4 {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.769049] env[63355]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2e0c00-a1fa-4e77-8a8d-68b365be778b {{(pid=63355) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.782137] env[63355]: DEBUG nova.compute.provider_tree [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed in ProviderTree for provider: 47abb610-db7e-4770-911d-187dd075ef8b {{(pid=63355) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1432.285152] env[63355]: DEBUG nova.scheduler.client.report [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Inventory has not changed for provider 47abb610-db7e-4770-911d-187dd075ef8b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63355) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1432.789935] env[63355]: DEBUG nova.compute.resource_tracker [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63355) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1432.790312] env[63355]: DEBUG oslo_concurrency.lockutils [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.095s {{(pid=63355) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1438.133634] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1438.134043] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1438.134043] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Starting heal instance info cache {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1439.144177] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Didn't find any instances for network info cache update. {{(pid=63355) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1439.144546] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1439.144658] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1439.144771] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1439.144919] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1439.145064] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63355) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1439.145225] env[63355]: DEBUG oslo_service.periodic_task [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63355) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1439.145345] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] Cleaning up deleted instances {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1439.650753] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] There are 12 instances to clean {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1439.650943] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: c2629183-3353-4bc6-8bf6-8fc72a874620] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1440.154676] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 9f7c978f-e993-4fa9-9402-489545a10020] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1440.657684] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 09e632ab-cd73-48cd-a4f3-308c247794c8] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1441.161161] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: e4c425fa-0b13-456c-b12b-c1ff18593536] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1441.664314] env[63355]: DEBUG nova.compute.manager [None req-d1b1a762-acb9-41ff-9fd7-3fe06118b415 None None] [instance: 5b246e6a-9974-49e4-b213-7664c90e8420] Instance has had 0 of 5 cleanup attempts {{(pid=63355) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}}